Augmented Lagrangian Methods

Size: px
Start display at page:

Download "Augmented Lagrangian Methods"

Transcription

1 Augmented Lagrangian Methods Mário A. T. Figueiredo 1 and Stephen J. Wright 2 1 Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal 2 Computer Sciences Department, University of Wisconsin, Madison, WI, USA HIM, Bonn, January 2016 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

2 Minimization with Linear Constraints: Basics Consider the linearly constrained problem, min f () s.t. A = b, where f : R n R is smooth. How do we recognize that a point is a solution of this problem? (Such optimality conditions can provide the foundation for algorithms.) Karush-Kuhn-Tucker (KKT) condition is a first-order necessary condition. If is a local solution, there eists a vector of Lagrange multipliers λ R m such that f ( ) = A T λ, A = b. When f is smooth and conve, these conditions are also sufficient. (In fact, it s enough for f to be conve on the null space of A.) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

3 Minimization with Linear Constraints Define the Lagrangian function: L(, λ) := f () + λ T (A b). Can write the KKT conditions in terms of L as follows: [ L(, λ L( ) =, λ ] ) λ L(, λ = 0. ) Suppose now that f is conve but not smooth. First-order optimality conditions (necessary and sufficient) are that there eists λ R m such that A T λ f ( ), A = b, where f is the subdifferential. In terms of the Lagrangian, we have 0 L(, λ ), λ L(, λ ) = 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

4 Augmented Lagrangian Methods With f proper, lower semi-continuous, and conve, consider: min f () s.t. A = b. The augmented Lagrangian is (with ρ > 0) L(, λ; ρ) := f () + λ T (A b) + ρ }{{} 2 A b 2 2. }{{} Lagrangian augmentation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

5 Augmented Lagrangian Methods With f proper, lower semi-continuous, and conve, consider: min f () s.t. A = b. The augmented Lagrangian is (with ρ > 0) L(, λ; ρ) := f () + λ T (A b) + ρ }{{} 2 A b 2 2. }{{} Lagrangian augmentation Basic augmented Lagrangian (a.k.a. method of multipliers) is (Hestenes, 1969; Powell, 1969) k = arg min L(, λ k 1 ; ρ); λ k = λ k 1 + ρ(a k b); M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

6 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

7 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. This equivalence is not very useful, computationally: the ma λ function is highly nonsmooth w.r.t.. Smooth it by adding a proimal point term, penalizing deviations from a prior estimate λ: min { ma f () + λ T (A b) 1 2 λ λ λ 2ρ }. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

8 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. This equivalence is not very useful, computationally: the ma λ function is highly nonsmooth w.r.t.. Smooth it by adding a proimal point term, penalizing deviations from a prior estimate λ: min { ma f () + λ T (A b) 1 2 λ λ λ 2ρ }. Maimization w.r.t. λ is now trivial (a concave quadratic), yielding λ = λ + ρ(a b). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

9 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

10 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). Hence can view the augmented Lagrangian process as: min L(, λ; ρ) to get new ; Shift the prior on λ by updating to the latest ma: λ + ρ(a b). repeat until convergence. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

11 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). Hence can view the augmented Lagrangian process as: min L(, λ; ρ) to get new ; Shift the prior on λ by updating to the latest ma: λ + ρ(a b). repeat until convergence. Add subscripts, and we recover the augmented Lagrangian algorithm of the first slide! Can also increase ρ (to sharpen the effect of the pro term), if needed. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

12 Inequality Constraints, Nonlinear Constraints The same derivation can be used for inequality constraints: min f () s.t. A b. Apply the same reasoning to the constrained min-ma formulation: min ma f () λ 0 λt (A b). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

13 Inequality Constraints, Nonlinear Constraints The same derivation can be used for inequality constraints: min f () s.t. A b. Apply the same reasoning to the constrained min-ma formulation: min ma f () λ 0 λt (A b). After the pro-term is added, can find the minimizing λ in closed form (as for pro-operators). Leads to update formula: ma ( λ + ρ(a b), 0 ). This derivation etends immediately to nonlinear constraints c() = 0 or c() 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

14 Eplicit Constraints, Inequality Constraints There may be other constraints on (such as Ω) that we prefer to handle eplicitly in the subproblem. For the formulation min f (), s.t. A = b, Ω, the min step can enforce Ω eplicitly: k = arg min Ω L(, λ k 1; ρ); λ k = λ k 1 + ρ(a k b); M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

15 Eplicit Constraints, Inequality Constraints There may be other constraints on (such as Ω) that we prefer to handle eplicitly in the subproblem. For the formulation min f (), s.t. A = b, Ω, the min step can enforce Ω eplicitly: k = arg min Ω L(, λ k 1; ρ); λ k = λ k 1 + ρ(a k b); This gives an alternative way to handle inequality constraints: introduce slacks s, and enforce them eplicitly. That is, replace by min f () s.t. c() 0, min f () s.t. c() = s, s 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

16 Eplicit Constraints, Inequality Constraints (Cont.) The augmented Lagrangian is now L(, s, λ; ρ) := f () + λ T (c() s) + ρ 2 c() s 2 2. Enforce s 0 eplicitly in the subproblem: ( k, s k ) = arg min,s L(, s, λ k 1; ρ), s.t. s 0; λ k = λ k 1 + ρ(c( k ) s k ) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

17 Eplicit Constraints, Inequality Constraints (Cont.) The augmented Lagrangian is now L(, s, λ; ρ) := f () + λ T (c() s) + ρ 2 c() s 2 2. Enforce s 0 eplicitly in the subproblem: ( k, s k ) = arg min,s L(, s, λ k 1; ρ), s.t. s 0; λ k = λ k 1 + ρ(c( k ) s k ) There are good algorithmic options for dealing with bound constraints s 0 (gradient projection and its enhancements). This is used in the Lancelot code (Conn et al., 1992). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

18 Quick History of Augmented Lagrangian Dates from at least 1969: Hestenes, Powell. Developments in 1970s, early 1980s by Rockafellar, Bertsekas, and others. Lancelot code for nonlinear programming: Conn, Gould, Toint, around 1992 (Conn et al., 1992). Lost favor somewhat as an approach for general nonlinear programming during the net 15 years. Recent revival in the contet of sparse optimization and its many applications, in conjunction with splitting / coordinate descent. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

19 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

20 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. Standard AL would minimize L(, z, λ; ρ) w.r.t. (, z) jointly. However, since coupled in the quadratic term, separability is lost. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

21 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. Standard AL would minimize L(, z, λ; ρ) w.r.t. (, z) jointly. However, since coupled in the quadratic term, separability is lost. In ADMM, minimize over and z separately and sequentially: k = arg min L(, z k 1, λ k 1 ; ρ); z k = arg min L( k, z, λ k 1 ; ρ); z λ k = λ k 1 + ρ(a k + Bz k c). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

22 ADMM Main features of ADMM: Does one cycle of block-coordinate descent in (, z). The minimizations over and z add only a quadratic term to f and h, respectively. Usually does not alter the cost much. Can perform the (, z) minimizations ineactly. Can add eplicit (separated) constraints: Ω, z Ω z. Many (many!) recent applications to compressed sensing, image processing, matri completion, sparse principal components analysis... ADMM has a rich collection of antecendents, dating even to the 1950s (operator splitting). For an comprehensive recent survey, including a diverse collection of machine learning applications, see Boyd et al. (2011). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

23 ADMM for Consensus Optimization Given the unconstrained (but separable) problem m min f i (), form m copies of the, with the original as a master variable: m f i ( i ) subject to i = 0, i = 1, 2,..., m. min, 1, 2,..., m i=1 i=1 Apply ADMM, with z = ( 1, 2,..., m ). Get m L(, 1, 2,..., m, λ 1,..., λ m ; ρ) = f i ( i )+(λ i ) T ( i )+ ρ 2 i 2 2. i=1 The minimization w.r.t. z = ( 1, 2,..., m ) is separable! k i = arg min f i ( i )+(λ i i k 1 )T ( i k 1 )+ ρ k 2 i k 1 2 2, i = 1, 2,..., m. Can be implemented in parallel. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

24 Consensus, continued The minimization w.r.t. can be done eplicitly averaging: k = 1 m m i=1 ( k i + 1 ) λ i k 1. ρ k Update to λ i can also be done in parallel, once the new k is known (and communicated): λ i k = λi k 1 + ρ k( i k k), i = 1, 2,..., m. If the initial λ i 0 have m i=1 λi 0 =, can see that m iterations k. Can simplify the update for k : i=1 λi k = 0 at all k = 1 m m k i. i=1 Gather-Scatter implementation. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

25 ADMM for Awkward Intersections The feasible set is sometimes an intersection of two or more conve sets that are easy to handle separately (e.g. projections are easily computable), but whose intersection is more difficult to work with. Eample: Optimization over the cone of doubly nonnegative matrices: General form: min X f (X ) s.t. X 0, X 0. min f () s.t. Ω i, i = 1, 2,..., m Again, use a different copy i for each set, and constrain them all to be the same: min f () s.t. i Ω i, i = 0, i = 1, 2,..., m., 1, 2,..., m M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

26 ADMM for Awkward Intersections Separable minimizations over Ω i, i = 1, 2,..., m: i k = arg min i Ω i (λ i k 1 )T ( i k 1 ) + ρ k 2 k i 2 2, i = 1, 2,..., m. Optimize over the master variable (unconstrained, with quadratic added to f ): k = arg min f () + Update multipliers: m i=1 (λ i k 1 )T ( i k 1 ) + ρ k 2 i k 1 2 2, λ i k = λi k 1 + ρ k( k k i ), i = 1, 2,..., m. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

27 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

28 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

29 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). Updating z k is a proimity computation: z k = pro h/ρ ( A k 1 d k 1 ) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

30 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). Updating z k is a proimity computation: z k = pro h/ρ ( A k 1 d k 1 ) Updating k may be hard: if f is quadratic, involves matri inversion; if f is not quadratic, may be as hard as the original problem. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

31 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

32 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

33 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). As in IST/FBS/PGA, convergence is still guaranteed with ineactly solved subproblems, as long as the errors are absolutely summable. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

34 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). As in IST/FBS/PGA, convergence is still guaranteed with ineactly solved subproblems, as long as the errors are absolutely summable. The recent eplosion of interest in ADMM is clear in the citation records of the review paper of Boyd et al. (2011) (2800 and counting) and of the paper by Eckstein and Bertsekas (1992): M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

35 ADMM for a More General Problem Consider the problem min R n J g j (H (j) ), where H (j) R p j n, i=1 and g 1,..., g J are l.s.c proper conve fuctions. Map it into min f () + h(a) as follows (with p = p p J ): f () = 0 A = H (1). H (J) R p n, z (1) h : R p1+ +p J R, h. = z (J) J g j (z (j) ) j=1 This leads to a convenient version of ADMM. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

36 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min Az z k d k 2 2 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

37 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

38 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k. ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min u g 1 + ρ 2 u H(1) k 1 + d (1) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) k j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

39 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

40 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 d k = d k 1 A k + z k j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

41 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 d k = d k 1 A k + z k Key features: matrices are handled separately from the pro operators; the pro operators are decoupled (can be computed in parallel); requires a matri inversion (can be a curse or a blessing). (Afonso et al., 2010; Setzer et al., 2010; Combettes and Pesquet, 2011) j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

42 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

43 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

44 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

45 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

46 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

47 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

48 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

49 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

50 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

51 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

52 ADMM for Sparse Inverse Covariance Reformulate as ma log det(x ) X, S τ X 1, X 0 Subproblems are: ma log det(x ) X, S τ Z 1 s.t. X Z = 0. X 0 X k := arg ma X log det(x ) X, S U k 1, X Z k 1 ρ k 2 X Z k 1 2 F := arg ma log det(x ) X, S ρ k X 2 X Z k 1 + U k /ρ k 2 F Z k := pro τ/ρk (X k + U k ); U k+1 := U k + ρ k (X k Z k ). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

53 Solving for X Get optimality condition for the X subproblem by using X log det(x ) = X 1, when X is s.p.d. Thus, which is equivalent to Form eigendecomposition X 1 S ρ k (X Z k 1 + U k /ρ k ) = 0, X 1 ρ k X (S ρ k Z k 1 + U k ) = 0. (S ρ k Z k 1 + U k ) = QΛQ T, where Q is n n orthogonal and Λ is diagonal with elements λ i. Seek X with the form Q ΛQ T, where Λ has diagonals λ i. Must have 1 ρ k λ i λ i = 0, i = 1, 2,..., n. λ i Take positive roots: λ i = [λ i + λ 2 i + 4ρ k ]/(2ρ k ), i = 1, 2,..., n. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

54 References I Afonso, M., Bioucas-Dias, J., and Figueiredo, M. (2010). Fast image recovery using variable splitting and constrained optimization. IEEE Transactions on Image Processing, 19: Boyd, S., Parikh, N., Chu, E., Peleato, B., and Eckstein, J. (2011). Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, 3(1): Combettes, P. and Pesquet, J.-C. (2011). Signal recovery by proimal forward-backward splitting. In Fied-Point Algorithms for Inverse Problems in Science and Engineering, pages Springer. Conn, A., Gould, N., and Toint, P. (1992). LANCELOT: a Fortran package for large-scale nonlinear optimization (Release A). Springer Verlag, Heidelberg. Eckstein, J. and Bertsekas, D. (1992). On the Douglas-Rachford splitting method and the proimal point algorithm for maimal monotone operators. Mathematical Programming, 5: Hestenes, M. (1969). Multiplier and gradient methods. Journal of Optimization Theory and Applications, 4: Powell, M. (1969). A method for nonlinear constraints in minimization problems. In Fletcher, R., editor, Optimization, pages Academic Press, New York. Setzer, S., Steidl, G., and Teuber, T. (2010). Deblurring poissonian images by split bregman techniques. Journal of Visual Communication and Image Representation, 21: M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33

Augmented Lagrangian Methods

Augmented Lagrangian Methods Augmented Lagrangian Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Augmented Lagrangian IMA, August 2016 1 /

More information

Lecture 4 Duality and Decomposition Techniques

Lecture 4 Duality and Decomposition Techniques Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

Epigraph proximal algorithms for general convex programming

Epigraph proximal algorithms for general convex programming Epigraph proimal algorithms for general conve programming Matt Wytock, Po-Wei Wang and J. Zico Kolter Machine Learning Department Carnegie Mellon University mwytock@cs.cmu.edu Abstract This work aims at

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Constrained Optimization Marc Toussaint U Stuttgart Constrained Optimization General constrained optimization problem: Let R n, f : R n R, g : R n R m, h : R n R l find min

More information

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection Paul Armand 1 Ngoc Nguyen Tran 2 Institut de Recherche XLIM Université de Limoges Journées annuelles

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider

More information

Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers

Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Stephen Boyd MIIS Xi An, 1/7/12 source: Distributed Optimization and Statistical Learning via the Alternating

More information

DM6 Support Vector Machines

DM6 Support Vector Machines DM6 Support Vector Machines Outline Large margin linear classifier Linear separable Nonlinear separable Creating nonlinear classifiers: kernel trick Discussion on SVM Conclusion SVM: LARGE MARGIN LINEAR

More information

Optimization III: Constrained Optimization

Optimization III: Constrained Optimization Optimization III: Constrained Optimization CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization III: Constrained Optimization

More information

Unconstrained and Constrained Optimization

Unconstrained and Constrained Optimization Unconstrained and Constrained Optimization Agenda General Ideas of Optimization Interpreting the First Derivative Interpreting the Second Derivative Unconstrained Optimization Constrained Optimization

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

Alternating Direction Method of Multipliers

Alternating Direction Method of Multipliers Alternating Direction Method of Multipliers CS 584: Big Data Analytics Material adapted from Stephen Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_slides.pdf) & Ryan Tibshirani (http://stat.cmu.edu/~ryantibs/convexopt/lectures/21-dual-meth.pdf)

More information

CME307/MS&E311 Optimization Theory Summary

CME307/MS&E311 Optimization Theory Summary CME307/MS&E311 Optimization Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/

More information

CME307/MS&E311 Theory Summary

CME307/MS&E311 Theory Summary CME307/MS&E311 Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/

More information

A primal-dual framework for mixtures of regularizers

A primal-dual framework for mixtures of regularizers A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland

More information

Iterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development

Iterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development Iterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development Mário A. T. Figueiredo Instituto de Telecomunicações and Instituto Superior Técnico, Technical University of Lisbon PORTUGAL

More information

A Short SVM (Support Vector Machine) Tutorial

A Short SVM (Support Vector Machine) Tutorial A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange

More information

A More Efficient Approach to Large Scale Matrix Completion Problems

A More Efficient Approach to Large Scale Matrix Completion Problems A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem

More information

Computational Optimization. Constrained Optimization Algorithms

Computational Optimization. Constrained Optimization Algorithms Computational Optimization Constrained Optimization Algorithms Same basic algorithms Repeat Determine descent direction Determine step size Take a step Until Optimal But now must consider feasibility,

More information

A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland

A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile and Volkan Cevher LIONS - EPFL Lausanne - Switzerland ABSTRACT Effectively solving many

More information

Convex optimization algorithms for sparse and low-rank representations

Convex optimization algorithms for sparse and low-rank representations Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,

More information

Lecture 7: Support Vector Machine

Lecture 7: Support Vector Machine Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,

More information

Introduction to Constrained Optimization

Introduction to Constrained Optimization Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm

Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Manya V. Afonso, José M. Bioucas-Dias, Mario A. T. Figueiredo To cite this version: Manya V. Afonso, José M. Bioucas-Dias,

More information

Efficient JPEG decompression by the alternating direction method of multipliers

Efficient JPEG decompression by the alternating direction method of multipliers 06 3rd International Conference on Pattern Recognition ICPR Cancún Center, Cancún, Méico, December 4-8, 06 Efficient JPEG decompression by the alternating direction method of multipliers Michal Šorel and

More information

The flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R

The flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R Journal of Machine Learning Research 6 (205) 553-557 Submitted /2; Revised 3/4; Published 3/5 The flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R Xingguo Li Department

More information

California Institute of Technology Crash-Course on Convex Optimization Fall Ec 133 Guilherme Freitas

California Institute of Technology Crash-Course on Convex Optimization Fall Ec 133 Guilherme Freitas California Institute of Technology HSS Division Crash-Course on Convex Optimization Fall 2011-12 Ec 133 Guilherme Freitas In this text, we will study the following basic problem: maximize x C f(x) subject

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization

PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,

More information

Lecture Notes: Constraint Optimization

Lecture Notes: Constraint Optimization Lecture Notes: Constraint Optimization Gerhard Neumann January 6, 2015 1 Constraint Optimization Problems In constraint optimization we want to maximize a function f(x) under the constraints that g i (x)

More information

Optimal Control Techniques for Dynamic Walking

Optimal Control Techniques for Dynamic Walking Optimal Control Techniques for Dynamic Walking Optimization in Robotics & Biomechanics IWR, University of Heidelberg Presentation partly based on slides by Sebastian Sager, Moritz Diehl and Peter Riede

More information

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal

More information

A Note on Smoothing Mathematical Programs with Equilibrium Constraints

A Note on Smoothing Mathematical Programs with Equilibrium Constraints Applied Mathematical Sciences, Vol. 3, 2009, no. 39, 1943-1956 A Note on Smoothing Mathematical Programs with Equilibrium Constraints M. A. Tawhid Department of Mathematics and Statistics School of Advanced

More information

Least-Squares Minimization Under Constraints EPFL Technical Report #

Least-Squares Minimization Under Constraints EPFL Technical Report # Least-Squares Minimization Under Constraints EPFL Technical Report # 150790 P. Fua A. Varol R. Urtasun M. Salzmann IC-CVLab, EPFL IC-CVLab, EPFL TTI Chicago TTI Chicago Unconstrained Least-Squares minimization

More information

Parallel and Distributed Sparse Optimization Algorithms

Parallel and Distributed Sparse Optimization Algorithms Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed

More information

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous

More information

Convex Optimization and Machine Learning

Convex Optimization and Machine Learning Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction

More information

Fast Image Recovery Using Variable Splitting and Constrained Optimization

Fast Image Recovery Using Variable Splitting and Constrained Optimization SUBMITTED TO THE IEEE TRANSACTIONS ON IMAGE PROCESSING, 009 Fast Image Recovery Using Variable Splitting and Constrained Optimization Manya V. Afonso, José M. Bioucas-Dias, and Mário A. T. Figueiredo arxiv:090.4887v

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing

More information

Constrained Optimization COS 323

Constrained Optimization COS 323 Constrained Optimization COS 323 Last time Introduction to optimization objective function, variables, [constraints] 1-dimensional methods Golden section, discussion of error Newton s method Multi-dimensional

More information

Distributed Alternating Direction Method of Multipliers

Distributed Alternating Direction Method of Multipliers Distributed Alternating Direction Method of Multipliers Ermin Wei and Asuman Ozdaglar Abstract We consider a network of agents that are cooperatively solving a global unconstrained optimization problem,

More information

Computational Optimization. Constrained Optimization

Computational Optimization. Constrained Optimization Computational Optimization Constrained Optimization Easiest Problem Linear equality constraints min f( ) f R s.. t A b A R, b R n m n m Null Space Representation Let * be a feasible point, A*b. Any other

More information

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake

More information

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation Xingguo Li Tuo Zhao Xiaoming Yuan Han Liu Abstract This paper describes an R package named flare, which implements

More information

Convex Optimization. Lijun Zhang Modification of

Convex Optimization. Lijun Zhang   Modification of Convex Optimization Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Modification of http://stanford.edu/~boyd/cvxbook/bv_cvxslides.pdf Outline Introduction Convex Sets & Functions Convex Optimization

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling

DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS Yuao Cheng, Houfeng Huang, Gang Wu, Qing Ling Department of Automation, University of Science and Technology of China, Hefei, China ABSTRACT

More information

Distributed Optimization via ADMM

Distributed Optimization via ADMM Distributed Optimization via ADMM Zhimin Peng Dept. Computational and Applied Mathematics Rice University Houston, TX 77005 Aug. 15, 2012 Main Reference: Distributed Optimization and Statistical Learning

More information

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21 Convex Programs COMPSCI 371D Machine Learning COMPSCI 371D Machine Learning Convex Programs 1 / 21 Logistic Regression! Support Vector Machines Support Vector Machines (SVMs) and Convex Programs SVMs are

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,

More information

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods 1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization

More information

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction

More information

Intro to Linear Programming. The problem that we desire to address in this course is loosely stated below.

Intro to Linear Programming. The problem that we desire to address in this course is loosely stated below. . Introduction Intro to Linear Programming The problem that we desire to address in this course is loosely stated below. Given a number of generators make price-quantity offers to sell (each provides their

More information

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations Peter Englert Machine Learning and Robotics Lab Universität Stuttgart Germany

More information

Linear methods for supervised learning

Linear methods for supervised learning Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

Numerical Method in Optimization as a Multi-stage Decision Control System

Numerical Method in Optimization as a Multi-stage Decision Control System Numerical Method in Optimization as a Multi-stage Decision Control System B.S. GOH Institute of Mathematical Sciences University of Malaya 50603 Kuala Lumpur MLYSI gohoptimum@gmail.com bstract: - Numerical

More information

Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy

Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Paroma Varma Stanford University paroma@stanford.edu Abstract In microscopy, shot noise dominates image formation, which

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

Convex Optimization MLSS 2015

Convex Optimization MLSS 2015 Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i

More information

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained

More information

Multiplicative Noise Removal Using Variable Splitting and Constrained Optimization

Multiplicative Noise Removal Using Variable Splitting and Constrained Optimization TO APPEAR IN THE IEEE TRANSACTIONS ON IMAGE PROCESSING, 2010. 1 Multiplicative Noise Removal Using Variable Splitting and Constrained Optimization José M. Bioucas-Dias, Member, IEEE, Mário A. T. Figueiredo,

More information

Alternating Direction Optimization for Imaging Inverse Problems

Alternating Direction Optimization for Imaging Inverse Problems Alternating Direction Optimization for Imaging Inverse Problems Mário A. T. Figueiredo Instituto Superior Técnico, University of Lisbon, Portugal Instituto de Telecomunicações Lisbon, Portugal Joint work

More information

Characterizing Improving Directions Unconstrained Optimization

Characterizing Improving Directions Unconstrained Optimization Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not

More information

Preface. and Its Applications 81, ISBN , doi: / , Springer Science+Business Media New York, 2013.

Preface. and Its Applications 81, ISBN , doi: / , Springer Science+Business Media New York, 2013. Preface This book is for all those interested in using the GAMS technology for modeling and solving complex, large-scale, continuous nonlinear optimization problems or applications. Mainly, it is a continuation

More information

Programs. Introduction

Programs. Introduction 16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.

More information

SnapVX: A Network-Based Convex Optimization Solver

SnapVX: A Network-Based Convex Optimization Solver Journal of Machine Learning Research 18 (2017) 1-5 Submitted 9/15; Revised 10/16; Published 2/17 SnapVX: A Network-Based Convex Optimization Solver David Hallac Christopher Wong Steven Diamond Abhijit

More information

APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING

APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING Second Edition P. Venkataraman Rochester Institute of Technology WILEY JOHN WILEY & SONS, INC. CONTENTS PREFACE xiii 1 Introduction 1 1.1. Optimization Fundamentals

More information

Image Super-Resolution Reconstruction Based On L 1/2 Sparsity

Image Super-Resolution Reconstruction Based On L 1/2 Sparsity Buletin Teknik Elektro dan Informatika (Bulletin of Electrical Engineering and Informatics) Vol. 3, No. 3, September 4, pp. 55~6 ISSN: 89-39 55 Image Super-Resolution Reconstruction Based On L / Sparsity

More information

Projection onto the probability simplex: An efficient algorithm with a simple proof, and an application

Projection onto the probability simplex: An efficient algorithm with a simple proof, and an application Proection onto the probability simplex: An efficient algorithm with a simple proof, and an application Weiran Wang Miguel Á. Carreira-Perpiñán Electrical Engineering and Computer Science, University of

More information

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP): Linköping University Optimization TAOP3(0) Department of Mathematics Examination Oleg Burdakov of 30 October 03 Assignment Consider the following linear programming problem (LP): max z = x + x s.t. x x

More information

Support Vector Machines

Support Vector Machines Support Vector Machines SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions 6. Dealing

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

Support Vector Machines.

Support Vector Machines. Support Vector Machines srihari@buffalo.edu SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions

More information

An augmented Lagrangian approach to non-convex SAO using diagonal quadratic approximations

An augmented Lagrangian approach to non-convex SAO using diagonal quadratic approximations Struct Multidisc Optim DOI 10.1007/s00158-008-0304- BRIEF NOTE An augmented Lagrangian approach to non-conve SAO using diagonal quadratic approimations Albert A. Groenwold L. F. P. Etman Schalk Kok Derren

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers Mid Year Review: AMSC/CMSC 663 and 664 The Alternating Direction Method of Multipliers An Adaptive Step-size Software Library Peter Sutor, Jr. psutor@umd.edu Project Advisor Dr. Tom Goldstein tomg@cs.umd.edu

More information

Lec 11 Rate-Distortion Optimization (RDO) in Video Coding-I

Lec 11 Rate-Distortion Optimization (RDO) in Video Coding-I CS/EE 5590 / ENG 401 Special Topics (17804, 17815, 17803) Lec 11 Rate-Distortion Optimization (RDO) in Video Coding-I Zhu Li Course Web: http://l.web.umkc.edu/lizhu/teaching/2016sp.video-communication/main.html

More information

LINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION

LINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION LINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION Nantheera Anantrasirichai 1, Marco Allinovi 2, Wesley Hayes 3, David Bull 1, Alin Achim 1 1 Visual Information Lab, University

More information

Efficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers

Efficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers Efficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers Xiaojie Zhang Samsung R&D America, Dallas, Texas 758 Email: eric.zhang@samsung.com Paul H. Siegel University

More information

Unlabeled Data Classification by Support Vector Machines

Unlabeled Data Classification by Support Vector Machines Unlabeled Data Classification by Support Vector Machines Glenn Fung & Olvi L. Mangasarian University of Wisconsin Madison www.cs.wisc.edu/ olvi www.cs.wisc.edu/ gfung The General Problem Given: Points

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

Gate Sizing by Lagrangian Relaxation Revisited

Gate Sizing by Lagrangian Relaxation Revisited Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007

More information

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song

More information

Recent Developments in Model-based Derivative-free Optimization

Recent Developments in Model-based Derivative-free Optimization Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:

More information

STRUCTURAL & MULTIDISCIPLINARY OPTIMIZATION

STRUCTURAL & MULTIDISCIPLINARY OPTIMIZATION STRUCTURAL & MULTIDISCIPLINARY OPTIMIZATION Pierre DUYSINX Patricia TOSSINGS Department of Aerospace and Mechanical Engineering Academic year 2018-2019 1 Course objectives To become familiar with the introduction

More information

Department of Electrical and Computer Engineering

Department of Electrical and Computer Engineering LAGRANGIAN RELAXATION FOR GATE IMPLEMENTATION SELECTION Yi-Le Huang, Jiang Hu and Weiping Shi Department of Electrical and Computer Engineering Texas A&M University OUTLINE Introduction and motivation

More information

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,

More information

Chapter 3 Numerical Methods

Chapter 3 Numerical Methods Chapter 3 Numerical Methods Part 1 3.1 Linearization and Optimization of Functions of Vectors 1 Problem Notation 2 Outline 3.1.1 Linearization 3.1.2 Optimization of Objective Functions 3.1.3 Constrained

More information

POGS Proximal Operator Graph Solver

POGS Proximal Operator Graph Solver POGS Proximal Operator Graph Solver Chris Fougner June 4, 2014 1 Introduction As data sets continue to increase in size, it becomes increasingly important to design algorithms that scale with the size

More information

Selected Topics in Column Generation

Selected Topics in Column Generation Selected Topics in Column Generation February 1, 2007 Choosing a solver for the Master Solve in the dual space(kelly s method) by applying a cutting plane algorithm In the bundle method(lemarechal), a

More information

Local and Global Minimum

Local and Global Minimum Local and Global Minimum Stationary Point. From elementary calculus, a single variable function has a stationary point at if the derivative vanishes at, i.e., 0. Graphically, the slope of the function

More information