Augmented Lagrangian Methods
|
|
- Colleen Harrison
- 5 years ago
- Views:
Transcription
1 Augmented Lagrangian Methods Mário A. T. Figueiredo 1 and Stephen J. Wright 2 1 Instituto de Telecomunicações, Instituto Superior Técnico, Lisboa, Portugal 2 Computer Sciences Department, University of Wisconsin, Madison, WI, USA HIM, Bonn, January 2016 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
2 Minimization with Linear Constraints: Basics Consider the linearly constrained problem, min f () s.t. A = b, where f : R n R is smooth. How do we recognize that a point is a solution of this problem? (Such optimality conditions can provide the foundation for algorithms.) Karush-Kuhn-Tucker (KKT) condition is a first-order necessary condition. If is a local solution, there eists a vector of Lagrange multipliers λ R m such that f ( ) = A T λ, A = b. When f is smooth and conve, these conditions are also sufficient. (In fact, it s enough for f to be conve on the null space of A.) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
3 Minimization with Linear Constraints Define the Lagrangian function: L(, λ) := f () + λ T (A b). Can write the KKT conditions in terms of L as follows: [ L(, λ L( ) =, λ ] ) λ L(, λ = 0. ) Suppose now that f is conve but not smooth. First-order optimality conditions (necessary and sufficient) are that there eists λ R m such that A T λ f ( ), A = b, where f is the subdifferential. In terms of the Lagrangian, we have 0 L(, λ ), λ L(, λ ) = 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
4 Augmented Lagrangian Methods With f proper, lower semi-continuous, and conve, consider: min f () s.t. A = b. The augmented Lagrangian is (with ρ > 0) L(, λ; ρ) := f () + λ T (A b) + ρ }{{} 2 A b 2 2. }{{} Lagrangian augmentation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
5 Augmented Lagrangian Methods With f proper, lower semi-continuous, and conve, consider: min f () s.t. A = b. The augmented Lagrangian is (with ρ > 0) L(, λ; ρ) := f () + λ T (A b) + ρ }{{} 2 A b 2 2. }{{} Lagrangian augmentation Basic augmented Lagrangian (a.k.a. method of multipliers) is (Hestenes, 1969; Powell, 1969) k = arg min L(, λ k 1 ; ρ); λ k = λ k 1 + ρ(a k b); M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
6 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
7 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. This equivalence is not very useful, computationally: the ma λ function is highly nonsmooth w.r.t.. Smooth it by adding a proimal point term, penalizing deviations from a prior estimate λ: min { ma f () + λ T (A b) 1 2 λ λ λ 2ρ }. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
8 A Favorite Derivation...more or less rigorous for conve f. Write the problem as min ma λ f () + λ T (A b). Obviously, the ma w.r.t. λ will be +, unless A = b, so this is equivalent to the original problem. This equivalence is not very useful, computationally: the ma λ function is highly nonsmooth w.r.t.. Smooth it by adding a proimal point term, penalizing deviations from a prior estimate λ: min { ma f () + λ T (A b) 1 2 λ λ λ 2ρ }. Maimization w.r.t. λ is now trivial (a concave quadratic), yielding λ = λ + ρ(a b). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
9 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
10 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). Hence can view the augmented Lagrangian process as: min L(, λ; ρ) to get new ; Shift the prior on λ by updating to the latest ma: λ + ρ(a b). repeat until convergence. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
11 A Favorite Derivation (Cont.) Inserting λ = λ + ρ(a b) leads to min f () + λ T (A b) + ρ 2 A b 2 = L(, λ; ρ). Hence can view the augmented Lagrangian process as: min L(, λ; ρ) to get new ; Shift the prior on λ by updating to the latest ma: λ + ρ(a b). repeat until convergence. Add subscripts, and we recover the augmented Lagrangian algorithm of the first slide! Can also increase ρ (to sharpen the effect of the pro term), if needed. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
12 Inequality Constraints, Nonlinear Constraints The same derivation can be used for inequality constraints: min f () s.t. A b. Apply the same reasoning to the constrained min-ma formulation: min ma f () λ 0 λt (A b). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
13 Inequality Constraints, Nonlinear Constraints The same derivation can be used for inequality constraints: min f () s.t. A b. Apply the same reasoning to the constrained min-ma formulation: min ma f () λ 0 λt (A b). After the pro-term is added, can find the minimizing λ in closed form (as for pro-operators). Leads to update formula: ma ( λ + ρ(a b), 0 ). This derivation etends immediately to nonlinear constraints c() = 0 or c() 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
14 Eplicit Constraints, Inequality Constraints There may be other constraints on (such as Ω) that we prefer to handle eplicitly in the subproblem. For the formulation min f (), s.t. A = b, Ω, the min step can enforce Ω eplicitly: k = arg min Ω L(, λ k 1; ρ); λ k = λ k 1 + ρ(a k b); M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
15 Eplicit Constraints, Inequality Constraints There may be other constraints on (such as Ω) that we prefer to handle eplicitly in the subproblem. For the formulation min f (), s.t. A = b, Ω, the min step can enforce Ω eplicitly: k = arg min Ω L(, λ k 1; ρ); λ k = λ k 1 + ρ(a k b); This gives an alternative way to handle inequality constraints: introduce slacks s, and enforce them eplicitly. That is, replace by min f () s.t. c() 0, min f () s.t. c() = s, s 0. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
16 Eplicit Constraints, Inequality Constraints (Cont.) The augmented Lagrangian is now L(, s, λ; ρ) := f () + λ T (c() s) + ρ 2 c() s 2 2. Enforce s 0 eplicitly in the subproblem: ( k, s k ) = arg min,s L(, s, λ k 1; ρ), s.t. s 0; λ k = λ k 1 + ρ(c( k ) s k ) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
17 Eplicit Constraints, Inequality Constraints (Cont.) The augmented Lagrangian is now L(, s, λ; ρ) := f () + λ T (c() s) + ρ 2 c() s 2 2. Enforce s 0 eplicitly in the subproblem: ( k, s k ) = arg min,s L(, s, λ k 1; ρ), s.t. s 0; λ k = λ k 1 + ρ(c( k ) s k ) There are good algorithmic options for dealing with bound constraints s 0 (gradient projection and its enhancements). This is used in the Lancelot code (Conn et al., 1992). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
18 Quick History of Augmented Lagrangian Dates from at least 1969: Hestenes, Powell. Developments in 1970s, early 1980s by Rockafellar, Bertsekas, and others. Lancelot code for nonlinear programming: Conn, Gould, Toint, around 1992 (Conn et al., 1992). Lost favor somewhat as an approach for general nonlinear programming during the net 15 years. Recent revival in the contet of sparse optimization and its many applications, in conjunction with splitting / coordinate descent. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
19 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
20 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. Standard AL would minimize L(, z, λ; ρ) w.r.t. (, z) jointly. However, since coupled in the quadratic term, separability is lost. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
21 Alternating Direction Method of Multipliers (ADMM) Consider now problems with a separable objective of the form min f () + h(z) s.t. A + Bz = c, (,z) for which the augmented Lagrangian is L(, z, λ; ρ) := f () + h(z) + λ T (A + Bz c) + ρ 2 A Bz c 2 2. Standard AL would minimize L(, z, λ; ρ) w.r.t. (, z) jointly. However, since coupled in the quadratic term, separability is lost. In ADMM, minimize over and z separately and sequentially: k = arg min L(, z k 1, λ k 1 ; ρ); z k = arg min L( k, z, λ k 1 ; ρ); z λ k = λ k 1 + ρ(a k + Bz k c). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
22 ADMM Main features of ADMM: Does one cycle of block-coordinate descent in (, z). The minimizations over and z add only a quadratic term to f and h, respectively. Usually does not alter the cost much. Can perform the (, z) minimizations ineactly. Can add eplicit (separated) constraints: Ω, z Ω z. Many (many!) recent applications to compressed sensing, image processing, matri completion, sparse principal components analysis... ADMM has a rich collection of antecendents, dating even to the 1950s (operator splitting). For an comprehensive recent survey, including a diverse collection of machine learning applications, see Boyd et al. (2011). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
23 ADMM for Consensus Optimization Given the unconstrained (but separable) problem m min f i (), form m copies of the, with the original as a master variable: m f i ( i ) subject to i = 0, i = 1, 2,..., m. min, 1, 2,..., m i=1 i=1 Apply ADMM, with z = ( 1, 2,..., m ). Get m L(, 1, 2,..., m, λ 1,..., λ m ; ρ) = f i ( i )+(λ i ) T ( i )+ ρ 2 i 2 2. i=1 The minimization w.r.t. z = ( 1, 2,..., m ) is separable! k i = arg min f i ( i )+(λ i i k 1 )T ( i k 1 )+ ρ k 2 i k 1 2 2, i = 1, 2,..., m. Can be implemented in parallel. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
24 Consensus, continued The minimization w.r.t. can be done eplicitly averaging: k = 1 m m i=1 ( k i + 1 ) λ i k 1. ρ k Update to λ i can also be done in parallel, once the new k is known (and communicated): λ i k = λi k 1 + ρ k( i k k), i = 1, 2,..., m. If the initial λ i 0 have m i=1 λi 0 =, can see that m iterations k. Can simplify the update for k : i=1 λi k = 0 at all k = 1 m m k i. i=1 Gather-Scatter implementation. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
25 ADMM for Awkward Intersections The feasible set is sometimes an intersection of two or more conve sets that are easy to handle separately (e.g. projections are easily computable), but whose intersection is more difficult to work with. Eample: Optimization over the cone of doubly nonnegative matrices: General form: min X f (X ) s.t. X 0, X 0. min f () s.t. Ω i, i = 1, 2,..., m Again, use a different copy i for each set, and constrain them all to be the same: min f () s.t. i Ω i, i = 0, i = 1, 2,..., m., 1, 2,..., m M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
26 ADMM for Awkward Intersections Separable minimizations over Ω i, i = 1, 2,..., m: i k = arg min i Ω i (λ i k 1 )T ( i k 1 ) + ρ k 2 k i 2 2, i = 1, 2,..., m. Optimize over the master variable (unconstrained, with quadratic added to f ): k = arg min f () + Update multipliers: m i=1 (λ i k 1 )T ( i k 1 ) + ρ k 2 i k 1 2 2, λ i k = λi k 1 + ρ k( k k i ), i = 1, 2,..., m. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
27 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
28 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
29 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). Updating z k is a proimity computation: z k = pro h/ρ ( A k 1 d k 1 ) M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
30 ADMM: A Simpler Form Often, a simpler version is enough: min f () + h(z) s.t. A = z, (,z) equivalent to min f () + h(a), often the one of interest. In this case, the ADMM can be written as k = arg min f () + ρ 2 A z k 1 d k z k = arg min z h(z) + ρ 2 A k 1 z d k d k = d k 1 (A k z k ) the so-called scaled version (Boyd et al., 2011). Updating z k is a proimity computation: z k = pro h/ρ ( A k 1 d k 1 ) Updating k may be hard: if f is quadratic, involves matri inversion; if f is not quadratic, may be as hard as the original problem. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
31 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
32 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
33 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). As in IST/FBS/PGA, convergence is still guaranteed with ineactly solved subproblems, as long as the errors are absolutely summable. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
34 ADMM: Convergence Consider the problem min f () + h(a), where f and h are lower semi-continuous, proper, conve functions and A has full column rank. The ADMM algorithm presented in the previous slide converges (for any ρ > 0) to a solution, if one eists, otherwise it diverges. This is a cornerstone result by Eckstein and Bertsekas (1992). As in IST/FBS/PGA, convergence is still guaranteed with ineactly solved subproblems, as long as the errors are absolutely summable. The recent eplosion of interest in ADMM is clear in the citation records of the review paper of Boyd et al. (2011) (2800 and counting) and of the paper by Eckstein and Bertsekas (1992): M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
35 ADMM for a More General Problem Consider the problem min R n J g j (H (j) ), where H (j) R p j n, i=1 and g 1,..., g J are l.s.c proper conve fuctions. Map it into min f () + h(a) as follows (with p = p p J ): f () = 0 A = H (1). H (J) R p n, z (1) h : R p1+ +p J R, h. = z (J) J g j (z (j) ) j=1 This leads to a convenient version of ADMM. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
36 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min Az z k d k 2 2 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
37 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
38 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k. ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min u g 1 + ρ 2 u H(1) k 1 + d (1) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) k j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
39 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
40 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 d k = d k 1 A k + z k j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
41 ADMM for a More General Problem (Cont.) Resulting instance of k = arg min z (1) k z (J) k ( J ) Az z k d k 2 1 ( J ) 2 = (H (j) ) T H (j) (H (j) ) T (z (j) k 1 + d (j) k 1 ) j=1 = arg min g 1 + ρ u 2 u H(1) k 1 + d (1) ( k = pro g1 /ρ H (1) k 1 d (1) ) k = arg min g J + ρ u 2 u H(J) k 1 + d (J) ( k = pro gj /ρ H (J) k 1 d (J) ) k 1 d k = d k 1 A k + z k Key features: matrices are handled separately from the pro operators; the pro operators are decoupled (can be computed in parallel); requires a matri inversion (can be a curse or a blessing). (Afonso et al., 2010; Setzer et al., 2010; Combettes and Pesquet, 2011) j=1 M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
42 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
43 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
44 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
45 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
46 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
47 Eample: Image Restoration using SALSA M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
48 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
49 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
50 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
51 ADMM for the Morozov Formulation M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
52 ADMM for Sparse Inverse Covariance Reformulate as ma log det(x ) X, S τ X 1, X 0 Subproblems are: ma log det(x ) X, S τ Z 1 s.t. X Z = 0. X 0 X k := arg ma X log det(x ) X, S U k 1, X Z k 1 ρ k 2 X Z k 1 2 F := arg ma log det(x ) X, S ρ k X 2 X Z k 1 + U k /ρ k 2 F Z k := pro τ/ρk (X k + U k ); U k+1 := U k + ρ k (X k Z k ). M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
53 Solving for X Get optimality condition for the X subproblem by using X log det(x ) = X 1, when X is s.p.d. Thus, which is equivalent to Form eigendecomposition X 1 S ρ k (X Z k 1 + U k /ρ k ) = 0, X 1 ρ k X (S ρ k Z k 1 + U k ) = 0. (S ρ k Z k 1 + U k ) = QΛQ T, where Q is n n orthogonal and Λ is diagonal with elements λ i. Seek X with the form Q ΛQ T, where Λ has diagonals λ i. Must have 1 ρ k λ i λ i = 0, i = 1, 2,..., n. λ i Take positive roots: λ i = [λ i + λ 2 i + 4ρ k ]/(2ρ k ), i = 1, 2,..., n. M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
54 References I Afonso, M., Bioucas-Dias, J., and Figueiredo, M. (2010). Fast image recovery using variable splitting and constrained optimization. IEEE Transactions on Image Processing, 19: Boyd, S., Parikh, N., Chu, E., Peleato, B., and Eckstein, J. (2011). Distributed optimization and statistical learning via the alternating direction method of multipliers. Foundations and Trends in Machine Learning, 3(1): Combettes, P. and Pesquet, J.-C. (2011). Signal recovery by proimal forward-backward splitting. In Fied-Point Algorithms for Inverse Problems in Science and Engineering, pages Springer. Conn, A., Gould, N., and Toint, P. (1992). LANCELOT: a Fortran package for large-scale nonlinear optimization (Release A). Springer Verlag, Heidelberg. Eckstein, J. and Bertsekas, D. (1992). On the Douglas-Rachford splitting method and the proimal point algorithm for maimal monotone operators. Mathematical Programming, 5: Hestenes, M. (1969). Multiplier and gradient methods. Journal of Optimization Theory and Applications, 4: Powell, M. (1969). A method for nonlinear constraints in minimization problems. In Fletcher, R., editor, Optimization, pages Academic Press, New York. Setzer, S., Steidl, G., and Teuber, T. (2010). Deblurring poissonian images by split bregman techniques. Journal of Visual Communication and Image Representation, 21: M. Figueiredo and S. Wright () Augmented Lagrangian Methods HIM, January / 33
Augmented Lagrangian Methods
Augmented Lagrangian Methods Stephen J. Wright 1 2 Computer Sciences Department, University of Wisconsin-Madison. IMA, August 2016 Stephen Wright (UW-Madison) Augmented Lagrangian IMA, August 2016 1 /
More informationLecture 4 Duality and Decomposition Techniques
Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More informationEpigraph proximal algorithms for general convex programming
Epigraph proimal algorithms for general conve programming Matt Wytock, Po-Wei Wang and J. Zico Kolter Machine Learning Department Carnegie Mellon University mwytock@cs.cmu.edu Abstract This work aims at
More informationIntroduction to Optimization
Introduction to Optimization Constrained Optimization Marc Toussaint U Stuttgart Constrained Optimization General constrained optimization problem: Let R n, f : R n R, g : R n R m, h : R n R l find min
More informationAn augmented Lagrangian method for equality constrained optimization with fast infeasibility detection
An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection Paul Armand 1 Ngoc Nguyen Tran 2 Institut de Recherche XLIM Université de Limoges Journées annuelles
More informationThe Alternating Direction Method of Multipliers
The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider
More informationDistributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers
Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Stephen Boyd MIIS Xi An, 1/7/12 source: Distributed Optimization and Statistical Learning via the Alternating
More informationDM6 Support Vector Machines
DM6 Support Vector Machines Outline Large margin linear classifier Linear separable Nonlinear separable Creating nonlinear classifiers: kernel trick Discussion on SVM Conclusion SVM: LARGE MARGIN LINEAR
More informationOptimization III: Constrained Optimization
Optimization III: Constrained Optimization CS 205A: Mathematical Methods for Robotics, Vision, and Graphics Doug James (and Justin Solomon) CS 205A: Mathematical Methods Optimization III: Constrained Optimization
More informationUnconstrained and Constrained Optimization
Unconstrained and Constrained Optimization Agenda General Ideas of Optimization Interpreting the First Derivative Interpreting the Second Derivative Unconstrained Optimization Constrained Optimization
More informationOptimization for Machine Learning
Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html
More informationAlternating Direction Method of Multipliers
Alternating Direction Method of Multipliers CS 584: Big Data Analytics Material adapted from Stephen Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_slides.pdf) & Ryan Tibshirani (http://stat.cmu.edu/~ryantibs/convexopt/lectures/21-dual-meth.pdf)
More informationCME307/MS&E311 Optimization Theory Summary
CME307/MS&E311 Optimization Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/
More informationCME307/MS&E311 Theory Summary
CME307/MS&E311 Theory Summary Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/~yyye http://www.stanford.edu/class/msande311/
More informationA primal-dual framework for mixtures of regularizers
A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland
More informationIterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development
Iterative Shrinkage/Thresholding g Algorithms: Some History and Recent Development Mário A. T. Figueiredo Instituto de Telecomunicações and Instituto Superior Técnico, Technical University of Lisbon PORTUGAL
More informationA Short SVM (Support Vector Machine) Tutorial
A Short SVM (Support Vector Machine) Tutorial j.p.lewis CGIT Lab / IMSC U. Southern California version 0.zz dec 004 This tutorial assumes you are familiar with linear algebra and equality-constrained optimization/lagrange
More informationA More Efficient Approach to Large Scale Matrix Completion Problems
A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem
More informationComputational Optimization. Constrained Optimization Algorithms
Computational Optimization Constrained Optimization Algorithms Same basic algorithms Repeat Determine descent direction Determine step size Take a step Until Optimal But now must consider feasibility,
More informationA PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS. LIONS - EPFL Lausanne - Switzerland
A PRIMAL-DUAL FRAMEWORK FOR MIXTURES OF REGULARIZERS Baran Gözcü, Luca Baldassarre, Quoc Tran-Dinh, Cosimo Aprile and Volkan Cevher LIONS - EPFL Lausanne - Switzerland ABSTRACT Effectively solving many
More informationConvex optimization algorithms for sparse and low-rank representations
Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,
More informationLecture 7: Support Vector Machine
Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each
More informationNonlinear Programming
Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,
More informationIntroduction to Constrained Optimization
Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR
More informationCOMS 4771 Support Vector Machines. Nakul Verma
COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron
More informationComparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles
INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.
More informationImage Restoration with Compound Regularization Using a Bregman Iterative Algorithm
Image Restoration with Compound Regularization Using a Bregman Iterative Algorithm Manya V. Afonso, José M. Bioucas-Dias, Mario A. T. Figueiredo To cite this version: Manya V. Afonso, José M. Bioucas-Dias,
More informationEfficient JPEG decompression by the alternating direction method of multipliers
06 3rd International Conference on Pattern Recognition ICPR Cancún Center, Cancún, Méico, December 4-8, 06 Efficient JPEG decompression by the alternating direction method of multipliers Michal Šorel and
More informationThe flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R
Journal of Machine Learning Research 6 (205) 553-557 Submitted /2; Revised 3/4; Published 3/5 The flare Package for High Dimensional Linear Regression and Precision Matrix Estimation in R Xingguo Li Department
More informationCalifornia Institute of Technology Crash-Course on Convex Optimization Fall Ec 133 Guilherme Freitas
California Institute of Technology HSS Division Crash-Course on Convex Optimization Fall 2011-12 Ec 133 Guilherme Freitas In this text, we will study the following basic problem: maximize x C f(x) subject
More informationDavid G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer
David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms
More informationPIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization
PIPA: A New Proximal Interior Point Algorithm for Large-Scale Convex Optimization Marie-Caroline Corbineau 1, Emilie Chouzenoux 1,2, Jean-Christophe Pesquet 1 1 CVN, CentraleSupélec, Université Paris-Saclay,
More informationLecture Notes: Constraint Optimization
Lecture Notes: Constraint Optimization Gerhard Neumann January 6, 2015 1 Constraint Optimization Problems In constraint optimization we want to maximize a function f(x) under the constraints that g i (x)
More informationOptimal Control Techniques for Dynamic Walking
Optimal Control Techniques for Dynamic Walking Optimization in Robotics & Biomechanics IWR, University of Heidelberg Presentation partly based on slides by Sebastian Sager, Moritz Diehl and Peter Riede
More informationSparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual
Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal
More informationA Note on Smoothing Mathematical Programs with Equilibrium Constraints
Applied Mathematical Sciences, Vol. 3, 2009, no. 39, 1943-1956 A Note on Smoothing Mathematical Programs with Equilibrium Constraints M. A. Tawhid Department of Mathematics and Statistics School of Advanced
More informationLeast-Squares Minimization Under Constraints EPFL Technical Report #
Least-Squares Minimization Under Constraints EPFL Technical Report # 150790 P. Fua A. Varol R. Urtasun M. Salzmann IC-CVLab, EPFL IC-CVLab, EPFL TTI Chicago TTI Chicago Unconstrained Least-Squares minimization
More informationParallel and Distributed Sparse Optimization Algorithms
Parallel and Distributed Sparse Optimization Algorithms Part I Ruoyu Li 1 1 Department of Computer Science and Engineering University of Texas at Arlington March 19, 2015 Ruoyu Li (UTA) Parallel and Distributed
More informationConic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding
Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationFast Image Recovery Using Variable Splitting and Constrained Optimization
SUBMITTED TO THE IEEE TRANSACTIONS ON IMAGE PROCESSING, 009 Fast Image Recovery Using Variable Splitting and Constrained Optimization Manya V. Afonso, José M. Bioucas-Dias, and Mário A. T. Figueiredo arxiv:090.4887v
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview 1. Overview of SVMs 2. Margin Geometry 3. SVM Optimization 4. Overlapping Distributions 5. Relationship to Logistic Regression 6. Dealing
More informationConstrained Optimization COS 323
Constrained Optimization COS 323 Last time Introduction to optimization objective function, variables, [constraints] 1-dimensional methods Golden section, discussion of error Newton s method Multi-dimensional
More informationDistributed Alternating Direction Method of Multipliers
Distributed Alternating Direction Method of Multipliers Ermin Wei and Asuman Ozdaglar Abstract We consider a network of agents that are cooperatively solving a global unconstrained optimization problem,
More informationComputational Optimization. Constrained Optimization
Computational Optimization Constrained Optimization Easiest Problem Linear equality constraints min f( ) f R s.. t A b A R, b R n m n m Null Space Representation Let * be a feasible point, A*b. Any other
More informationSupport Vector Machines. James McInerney Adapted from slides by Nakul Verma
Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake
More informationAn R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation
An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation Xingguo Li Tuo Zhao Xiaoming Yuan Han Liu Abstract This paper describes an R package named flare, which implements
More informationConvex Optimization. Lijun Zhang Modification of
Convex Optimization Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Modification of http://stanford.edu/~boyd/cvxbook/bv_cvxslides.pdf Outline Introduction Convex Sets & Functions Convex Optimization
More informationContents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.
page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5
More informationDISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling
DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS Yuao Cheng, Houfeng Huang, Gang Wu, Qing Ling Department of Automation, University of Science and Technology of China, Hefei, China ABSTRACT
More informationDistributed Optimization via ADMM
Distributed Optimization via ADMM Zhimin Peng Dept. Computational and Applied Mathematics Rice University Houston, TX 77005 Aug. 15, 2012 Main Reference: Distributed Optimization and Statistical Learning
More informationConvex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21
Convex Programs COMPSCI 371D Machine Learning COMPSCI 371D Machine Learning Convex Programs 1 / 21 Logistic Regression! Support Vector Machines Support Vector Machines (SVMs) and Convex Programs SVMs are
More informationIntroduction to Optimization
Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,
More informationUnconstrained Optimization Principles of Unconstrained Optimization Search Methods
1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization
More informationLECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach
LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction
More informationIntro to Linear Programming. The problem that we desire to address in this course is loosely stated below.
. Introduction Intro to Linear Programming The problem that we desire to address in this course is loosely stated below. Given a number of generators make price-quantity offers to sell (each provides their
More informationInverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations
Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations Peter Englert Machine Learning and Robotics Lab Universität Stuttgart Germany
More informationLinear methods for supervised learning
Linear methods for supervised learning LDA Logistic regression Naïve Bayes PLA Maximum margin hyperplanes Soft-margin hyperplanes Least squares resgression Ridge regression Nonlinear feature maps Sometimes
More informationLecture 2 September 3
EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give
More informationNumerical Method in Optimization as a Multi-stage Decision Control System
Numerical Method in Optimization as a Multi-stage Decision Control System B.S. GOH Institute of Mathematical Sciences University of Malaya 50603 Kuala Lumpur MLYSI gohoptimum@gmail.com bstract: - Numerical
More informationApplication of Proximal Algorithms to Three Dimensional Deconvolution Microscopy
Application of Proximal Algorithms to Three Dimensional Deconvolution Microscopy Paroma Varma Stanford University paroma@stanford.edu Abstract In microscopy, shot noise dominates image formation, which
More informationSection 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1
Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients
More informationConvex Optimization MLSS 2015
Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i
More informationPerformance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization
6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained
More informationMultiplicative Noise Removal Using Variable Splitting and Constrained Optimization
TO APPEAR IN THE IEEE TRANSACTIONS ON IMAGE PROCESSING, 2010. 1 Multiplicative Noise Removal Using Variable Splitting and Constrained Optimization José M. Bioucas-Dias, Member, IEEE, Mário A. T. Figueiredo,
More informationAlternating Direction Optimization for Imaging Inverse Problems
Alternating Direction Optimization for Imaging Inverse Problems Mário A. T. Figueiredo Instituto Superior Técnico, University of Lisbon, Portugal Instituto de Telecomunicações Lisbon, Portugal Joint work
More informationCharacterizing Improving Directions Unconstrained Optimization
Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not
More informationPreface. and Its Applications 81, ISBN , doi: / , Springer Science+Business Media New York, 2013.
Preface This book is for all those interested in using the GAMS technology for modeling and solving complex, large-scale, continuous nonlinear optimization problems or applications. Mainly, it is a continuation
More informationPrograms. Introduction
16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen
More informationDetecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationSurrogate Gradient Algorithm for Lagrangian Relaxation 1,2
Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.
More informationSnapVX: A Network-Based Convex Optimization Solver
Journal of Machine Learning Research 18 (2017) 1-5 Submitted 9/15; Revised 10/16; Published 2/17 SnapVX: A Network-Based Convex Optimization Solver David Hallac Christopher Wong Steven Diamond Abhijit
More informationAPPLIED OPTIMIZATION WITH MATLAB PROGRAMMING
APPLIED OPTIMIZATION WITH MATLAB PROGRAMMING Second Edition P. Venkataraman Rochester Institute of Technology WILEY JOHN WILEY & SONS, INC. CONTENTS PREFACE xiii 1 Introduction 1 1.1. Optimization Fundamentals
More informationImage Super-Resolution Reconstruction Based On L 1/2 Sparsity
Buletin Teknik Elektro dan Informatika (Bulletin of Electrical Engineering and Informatics) Vol. 3, No. 3, September 4, pp. 55~6 ISSN: 89-39 55 Image Super-Resolution Reconstruction Based On L / Sparsity
More informationProjection onto the probability simplex: An efficient algorithm with a simple proof, and an application
Proection onto the probability simplex: An efficient algorithm with a simple proof, and an application Weiran Wang Miguel Á. Carreira-Perpiñán Electrical Engineering and Computer Science, University of
More informationDepartment of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):
Linköping University Optimization TAOP3(0) Department of Mathematics Examination Oleg Burdakov of 30 October 03 Assignment Consider the following linear programming problem (LP): max z = x + x s.t. x x
More informationSupport Vector Machines
Support Vector Machines SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions 6. Dealing
More informationPRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction
PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem
More informationSupport Vector Machines.
Support Vector Machines srihari@buffalo.edu SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions
More informationAn augmented Lagrangian approach to non-convex SAO using diagonal quadratic approximations
Struct Multidisc Optim DOI 10.1007/s00158-008-0304- BRIEF NOTE An augmented Lagrangian approach to non-conve SAO using diagonal quadratic approimations Albert A. Groenwold L. F. P. Etman Schalk Kok Derren
More informationThe Alternating Direction Method of Multipliers
Mid Year Review: AMSC/CMSC 663 and 664 The Alternating Direction Method of Multipliers An Adaptive Step-size Software Library Peter Sutor, Jr. psutor@umd.edu Project Advisor Dr. Tom Goldstein tomg@cs.umd.edu
More informationLec 11 Rate-Distortion Optimization (RDO) in Video Coding-I
CS/EE 5590 / ENG 401 Special Topics (17804, 17815, 17803) Lec 11 Rate-Distortion Optimization (RDO) in Video Coding-I Zhu Li Course Web: http://l.web.umkc.edu/lizhu/teaching/2016sp.video-communication/main.html
More informationLINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION
LINE DETECTION IN SPECKLE IMAGES USING RADON TRANSFORM AND l 1 REGULARIZATION Nantheera Anantrasirichai 1, Marco Allinovi 2, Wesley Hayes 3, David Bull 1, Alin Achim 1 1 Visual Information Lab, University
More informationEfficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers
Efficient Iterative LP Decoding of LDPC Codes with Alternating Direction Method of Multipliers Xiaojie Zhang Samsung R&D America, Dallas, Texas 758 Email: eric.zhang@samsung.com Paul H. Siegel University
More informationUnlabeled Data Classification by Support Vector Machines
Unlabeled Data Classification by Support Vector Machines Glenn Fung & Olvi L. Mangasarian University of Wisconsin Madison www.cs.wisc.edu/ olvi www.cs.wisc.edu/ gfung The General Problem Given: Points
More informationApplied Lagrange Duality for Constrained Optimization
Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity
More informationChapter II. Linear Programming
1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION
More informationGate Sizing by Lagrangian Relaxation Revisited
Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007
More informationA fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]
A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009] Yongjia Song University of Wisconsin-Madison April 22, 2010 Yongjia Song
More informationRecent Developments in Model-based Derivative-free Optimization
Recent Developments in Model-based Derivative-free Optimization Seppo Pulkkinen April 23, 2010 Introduction Problem definition The problem we are considering is a nonlinear optimization problem with constraints:
More informationSTRUCTURAL & MULTIDISCIPLINARY OPTIMIZATION
STRUCTURAL & MULTIDISCIPLINARY OPTIMIZATION Pierre DUYSINX Patricia TOSSINGS Department of Aerospace and Mechanical Engineering Academic year 2018-2019 1 Course objectives To become familiar with the introduction
More informationDepartment of Electrical and Computer Engineering
LAGRANGIAN RELAXATION FOR GATE IMPLEMENTATION SELECTION Yi-Le Huang, Jiang Hu and Weiping Shi Department of Electrical and Computer Engineering Texas A&M University OUTLINE Introduction and motivation
More informationAdvanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude
Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,
More informationChapter 3 Numerical Methods
Chapter 3 Numerical Methods Part 1 3.1 Linearization and Optimization of Functions of Vectors 1 Problem Notation 2 Outline 3.1.1 Linearization 3.1.2 Optimization of Objective Functions 3.1.3 Constrained
More informationPOGS Proximal Operator Graph Solver
POGS Proximal Operator Graph Solver Chris Fougner June 4, 2014 1 Introduction As data sets continue to increase in size, it becomes increasingly important to design algorithms that scale with the size
More informationSelected Topics in Column Generation
Selected Topics in Column Generation February 1, 2007 Choosing a solver for the Master Solve in the dual space(kelly s method) by applying a cutting plane algorithm In the bundle method(lemarechal), a
More informationLocal and Global Minimum
Local and Global Minimum Stationary Point. From elementary calculus, a single variable function has a stationary point at if the derivative vanishes at, i.e., 0. Graphically, the slope of the function
More information