A proximal center-based decomposition method for multi-agent convex optimization

Size: px
Start display at page:

Download "A proximal center-based decomposition method for multi-agent convex optimization"

Transcription

1 Proceedings of the 47th IEEE Conference on Decision and Control Cancun, Mexico, Dec. 9-11, 2008 A proximal center-based decomposition method for multi-agent convex optimization Ion Necoara and Johan A.K. Suykens Abstract In this paper we develop a new dual decomposition method for optimizing a sum of convex objective functions corresponding to multiple agents but with coupled constraints. In our method we define a smooth Lagrangian, by using a smoothing technique developed by Nesterov, which preserves separability of the problem. With this approach we propose a new decomposition method (the proximal center method) for which efficiency estimates are derived and which improves the bounds on the number of iterations of the classical dual gradient scheme by an order of magnitude. The method involves every agent optimizing an objective function that is the sum of his own objective function and a smoothing term while the coordination between agents is performed via the Lagrange multipliers corresponding to the coupled constraints. Applications of the new method for solving distributed model predictive control or network optimization problems are also illustrated. I. INTRODUCTION Recent advances in computer power have led to the derivation of many parallel and distributed computation methods for solving large-scale optimization problems (e.g. [1]). For separable convex problems, i.e. separable objective function but with coupled constraints (they arise in many fields of engineering: e.g. network optimization [9], [14], distributed model predictive control (MPC) [13]), many researchers have proposed dual decomposition algorithms such as the dual subgradient method [1], the alternating direction method [1], [4], [5], [12], proximal method of multipliers [2], partial inverse method [11], etc. In general, these methods are based on alternating minimization in a Gauss-Seidel fashion of an (Augmented) Lagrangian followed by a steepest ascent update for the multipliers. However, the stepsize parameter which has a very strong influence on the convergence rate of these methods is very difficult to tune and the minimization in a Gauss-Seidel fashion slows down the convergence towards an optimum. Moreover, they do not provide any complexity estimates for the general case (linear convergence is obtained e.g. for strongly convex functions). The purpose of this paper is to propose a new decomposition method for separable convex optimization problems that overcomes the disadvantages mentioned above. Based on a smoothing technique recently developed by Nesterov in [8], we obtain a smooth Lagrangian that preserves separability of the problem. Using this smooth Lagrangian, we derive a new dual decomposition method in which the corresponding parameters are selected optimally and thus straightforward to tune. In contrast to the dual gradient I. Necoara and J.A.K. Suykens are with the Katholieke Universiteit Leuven, Department of Electrical Engineering, ESAT SCD, Kasteelpark Arenberg 10, B 3001 Leuven (Heverlee), Belgium. {ion.necoara,johan.suykens}@esat.kuleuven.be update for the multipliers used by most of the decomposition methods from the literature, our method uses an optimal first-order oracle scheme (see e.g. [7], [8]) for updating the multipliers. Therefore, we derive for the new method an efficiency estimate for the general case which improves with one order of magnitude the complexity of the classical dual gradient method (i.e. the steepest ascent update). Up to our knowledge these are the first efficiency estimate results of a dual decomposition method for separable non-strongly convex programs. The new algorithm is suitable for solving distributed model predictive control problems or network optimization problems since it is highly parallelizable and thus it can be effectively implemented on parallel processors. The reminder of this paper is organized as follows. Section II contains the problem formulation, followed by a brief description of some of the dual decomposition methods derived in the literature. The main results of the paper are presented in Section III, where we describe our new decomposition method and its main properties. In Section IV we show that network optimization or distributed model predictive control problems corresponding to linear systems with interacting subsystem dynamics and decoupled costs can be recast in the framework of separable convex problems for which our algorithm can be applied. Finally, numerical simulations on some test problems are included. II. PRELIMINARIES A. Splitting methods for separable convex programs An important application of convex duality theory is in decomposition algorithms for solving large-scale problems but with special structure. This type of problems arises e.g. in the context of large-scale networks which consists of multiple agents with different objectives or in the context of distributive MPC for linear systems with coupled subsystem dynamics. The following separable convex program will be considered in this paper: f = min x X,z Z { φ1 (x) + φ 2 (z) : Ax + Bz = b }, (1) where φ 1 : R m R and φ 2 : R p R are continuous convex functions on X and Z, respectively, A is an n m matrix, B an n p matrix, and b is in R n. In this paper we do not assume φ 1 and/or φ 2 to be strongly convex or even smooth. Moreover, we assume that X R m and Z R p are compact convex sets. We use different norms on R n, R m and R p, not necessarily the Euclidean norms. However, for simplicity in notation we do not use indices to specify the norms on R n, R m and R p, since from the context it is clear in which Euclidean space we are and which norm we use /08/$ IEEE 3077

2 Remark 2.1 Note that the method developed in this paper can treat both coupled inequalities Ax + Bz b and/or equalities Ax+Bz=b (see [6]). Moreover, we can consider any finite sum of φ i s. However, for simplicity of the exposition we restrict ourselves to problems of the form (1). Let, denote a scalar product on the Euclidean space R n. By forming the Lagrangian corresponding to the linear constraints (with the Lagrange multipliers λ R n ), i.e. L 0 (x,z,λ) = φ 1 (x) + φ 2 (z) + λ,ax + Bz b, and using the dual decomposition method, one arrives at the following decomposition algorithm: Algorithm 2.2: ([1]) for k 0 do 1. given λ k, minimize the Lagrangian (x k+1,z k+1 ) = arg min x X,z Z L 0 (x,z,λ k ), or equivalently minimize over x and z independently x k+1 = arg min x X [φ 1(x) + λ k,ax ] z k+1 = arg min z Z [φ 2(z) + λ k,bz ] 2. update the multipliers by the iteration λ k+1 = λ k + c k (Ax k+1 + Bz k+1 b), where c k is a positive step-size. The following assumption holds throughout the paper: Assumption 2.3: Constraint qualification holds for (1). It can be shown that Algorithm 2.2 is convergent under Assumption 2.3 and the assumption that both φ 1 and φ 2 are strongly convex functions (the later guarantees that the minimizer (x k+1,z k+1 ) is unique). In fact, under the assumption of strong convexity it follows that the dual function f 0 (λ) = min φ 1(x) + φ 2 (z) + λ,ax + Bz b x X,z Z is differentiable [10], and thus Algorithm 2.2 is the gradient method with stepsize c k for maximizing the dual function. However, for many interesting problems, especially arising from transformations that leads to decomposition, the functions φ 1 and φ 2 are not strongly convex. There are some methods (e.g. alternating direction method [1], [5], [12], proximal point method [2], partial inverse method [11]) that overcome this difficulty based on alternating minimization in a Gauss-Seidel fashion of the Augmented Lagrangian, followed by a steepest ascent update of the multipliers. A computational drawback of these schemes is that the proxterm c 2 Ax+Bz b 2, using the Euclidean norm framework, present in the Augmented Lagrangian is not separable in x and z. Another disadvantage is that they cannot deal with coupled inequalities in general. Moreover, these schemes were shown to be very sensitive to the value of the parameter c, with difficulties in practice to obtain the best convergence rate. Some heuristics for choosing c can be found in the literature [2], [5], [12]. Note that alternating direction method variants which allow for inexact minimization was proposed in e.g. [2]. A closely related method is the partial inverse of a monotone operator developed in [11]. B. Accelerated scheme for smooth maximization In this section we briefly describe an accelerated scheme for smooth convex functions developed by Nesterov in [7], [8]. Let f be a given concave and differentiable function on a closed convex set ˆQ R n that has a Lipschitz continuous gradient with Lipschitz constant L. Definition 2.4: [8] We define a prox-function d of the set ˆQ as a function with the following properties: d is continuous, strongly convex on ˆQ with convexity parameter σ, and u 0 is the center of the set ˆQ, i.e. u 0 = arg min x ˆQ d(x) such that d(u 0 ) = 0. The goal is to find an approximate solution of the convex optimization problem x = arg max x ˆQ f(x). In Nesterov s scheme three sequences of points from ˆQ are updated recursively: {u k } k 0, {x k } k 0, and {v k } k 0. The algorithm can be described as follows: Algorithm 2.5: ([8]) for k 0 do 1. compute f(u k ) 2. find x k = arg max x ˆQ f(u k ),x u k L 2 x uk 2 and define x k = arg max x { xk,x k 1,u k } f(x) 3. find v k = arg max x ˆQ L σ d(x) + k l+1 2 f(ul ),x u l 4. set u k+1 = k+1 k+3 xk + 2 k+3 vk. The main property of the Algorithm 2.5 is the following relation [8]: (k + 1)(k + 2) 4 f(x k ) max L x ˆQ σ d(x)+ l [f(ul ) + f(u l ),x u l ]. (2) Using this property Nesterov proves in [8] that the efficiency L estimates of Algorithm 2.5 is of the order O( ǫ ), higher than the corresponding pure gradient method for the same smooth problem by an order of magnitude. III. A NEW DECOMPOSITION METHOD BASED ON SMOOTHING THE LAGRANGIAN In this section we propose a new method to smoothen the Lagrangian of (1), inspired from [8]. This smoothing technique preserves the separability of the problem and moreover the corresponding parameters are easy to tune. Since separability is preserved under this smoothing technique, we derive a new dual decomposition method in which the multipliers are updated according to Algorithm 2.5. Moreover, we obtain an efficiency estimate of the new method for the general case. Note that with our method we can treat both coupled inequalities (Ax+Bz b) and/or equalities (Ax+Bz = b). A. Smoothing the Lagrangian Let d X and d Z be two prox-functions for the compact convex sets X and Z, with convexity parameter σ X and σ Z, respectively. Denote x 0 = arg min x X d X (x), z 0 = arg min z Z d Z (z). Since X and Z are compact and d X and d Z are continuous, we can choose finite and positive constants D X max x X d X (x) and D Z 3078

3 max z Z d Z (z). We also introduce the following notation A = max λ =1, x =1 λ,ax. Note that A = max x =1 Ax and Ax A x x, where s = max x =1 s,x is the corresponding dual norm of the norm used on R n [10]. Similarly for B. Let us introduce the following function: f c (λ) = min φ 1(x) + φ 2 (z)+ λ,ax + Bz b + x X,z Z c ( d X (x) + d Z (z) ), (3) where c is a positive smoothness parameter that will be defined in the sequel. Note that the objective function in (3) is separable in x and z, i.e. f c (λ) = λ,b + min x X [φ 1(x) + λ,ax + c d X (x)]+ min [φ 2(z) + λ,bz + c d Z (z)]. (4) z Z Denote by x(λ) and z(λ) the optimal solution of the minimization problem in x and z, respectively. Function f c has the following smoothness properties: Theorem 3.1: [6] The function f c is concave and continuously differentiable at any λ R n. Moreover, its gradient f c (λ) = Ax(λ) + Bz(λ) b is Lipschitz continuous with Lipschitz constant L c = A 2 cσ X cσ Z. The following inequalities also hold: f c (λ) f 0 (λ) f c (λ) c(d X + D Z ) λ R n. (5) Proof: The proof follows from standard convex argument and can be found in [6]. B. A proximal center based decomposition method In this section we derive a new decomposition method based on the smoothing technique described in Section III- A. The new algorithm, called here the proximal center algorithm, has the nice feature that the coordination between agents involves the maximization of a smooth convex objective function (i.e. with Lipschitz continuous gradient). Moreover, the resource allocation stage consists in solving in parallel by the each agent of a minimization problem with strongly convex objective. The new method belongs to the class of two-level algorithms [3] and is particularly suitable for separable convex problems where the minimizations over x and z in (4) are easily carried out. Let us apply Nesterov s accelerated method described in Algorithm 2.5 to the concave function f c that has a Lipschitz continuous gradient: max f c(λ), (6) λ Q where Q is a given closed convex set in R n that contains at least one optimal multiplier λ Λ (here Λ denotes the set of optimal multipliers). Notice that Q R n in the presence of linear equalities (i.e. Ax+Bz b = 0), Q R n +, where R + denotes the set of nonnegative real numbers, in the presence of linear inequalities (i.e. Ax + Bz b 0), or Q (R n1 R n2 + ), where n 1 + n 2 = n, when both, equalities and inequalities are present. Note that according to Algorithm 2.5 we also need to choose a prox-function d Q for the set Q with the convexity parameter σ Q and center u 0. The proximal center algorithm can be described as follows: Algorithm 3.2: for k 0 do 1. given u k compute in parallel x k+1 = arg min x X [φ 1(x) + u k,ax + c d X (x)], z k+1 = arg min z Z [φ 2(z) + u k,bz + c d Z (z)]. 2. compute f c (u k ), f c (u k ) = Ax k+1 + Bz k+1 b 3. find λ k = arg max λ Q f c (u k ),λ u k Lc 2 λ u k 2 and define λ k = arg max λ { λk,λ k 1,u k } f c (λ) 4. find v k = arg max λ Q Lc σ Q d Q (λ) + k l+1 2 f c(u l ),λ u l 5. set u k+1 = k+1 k+3 λk + 2 k+3 vk. The proximal center algorithm is suitable for decomposition since it is highly parallelizable: the agents can solve their corresponding minimization problems in parallel. This is an advantage of our method compared to most of the alternating direction methods based on Gauss-Seidel iterations that do not have this feature. We now derive a lower bound for the value of the objective function which will be used frequently in the sequel: Lemma 3.3: [6] For any λ Λ and ˆx X,ẑ Z, the following lower bound on primal gap holds: [φ 1 (ˆx)+φ 2 (ẑ)] f Aˆx + Bẑ b,λ λ Aˆx + Bẑ b. The previous lemma shows that if Aˆx + Bẑ b ǫ c, then the primal gap is bounded: for all ˆλ Q ǫ c λ φ 1 (ˆx) + φ 2 (ẑ)] f (7) φ 1 (ˆx) + φ 2 (ẑ) f 0 (ˆλ). Therefore, if we are able to derive an upper bound ǫ for the duality gap and ǫ c for the coupling constraints for some given ˆλ and ˆx X,ẑ Z, then we conclude that (ˆx,ẑ) is an (ǫ,ǫ c )-solution for problem (1) (since in this case ǫ c λ φ 1 (ˆx) + φ 2 (ẑ) f ǫ for all λ Λ ). The next theorem derives an upper bound on the duality gap for our method. Theorem 3.4: Assume that there exists a closed convex set Q that contains a λ Λ. Then, after k iterations we obtain an approximate solution to the problem (1) (ˆx,ẑ) = k 2(l+1) (k+1)(k+2) (xl+1,z l+1 ) and ˆλ = λ k which satisfy the following duality gap: 0 [φ 1 (ˆx) + φ 2 (ẑ)] f 0 (ˆλ) c(d X +D Z ) [ 4L c max λ Q σ Q (k + 1) 2 d Q(λ)+ Aˆx + Bẑ b,λ ]. (8) Proof: For an arbitrary c, we have from the inequality (2) that after k iterations the Lagrange multiplier ˆλ satisfies the following relation: (k + 1)(k + 2) f c (ˆλ) max 4 L c d Q (λ)+ λ Q σ Q l [f c(u l ) + f c (u l ),λ u l ]. 3079

4 In view of the previous inequality we have: f c (ˆλ) max λ Q 4L c σ Q (k + 1) 2 d Q(λ)+ (k + 1)(k + 2) [f c(u l ) + f c (u l ),λ u l ]. Now, we replace f c (u l ) and f c (u l ) with the expressions given in step 2 of Algorithm 3.2: (k + 1)(k + 2) [f c(u l ) + f c (u l ),λ u l ] = (k + 1)(k + 2) [φ 1(x l+1 ) + φ 2 (z l+1 )+ c ( d X (x l+1 ) + d Z (z l+1 ) ) + Ax l+1 + Bz l+1 b,λ ] (k + 1)(k + 2) [φ 1(x l+1 ) + φ 2 (z l+1 )+ Ax l+1 + Bz l+1 b,λ ] φ 1 (ˆx) + φ 2 (ẑ) + Aˆx + Bẑ b,λ, λ Q. The first inequality follows from the definition of a proxfunction and the second inequality follows from the fact that φ 1 and φ 2 are convex functions. Using the last relation and (5) we derive the bound (8) on the duality gap. We now show how to construct the set Q and how to choose optimally the smoothness parameter c. In the next two sections we discuss two cases depending on the choices for Q and d Q. C. Efficiency estimates for compact Q Let D Q be a positive constant satisfying max d Q(λ) D Q. (9) λ Q Let us note that we can choose D Q finite whenever Q is compact. In this section we specialize the result of Theorem 3.4 for the case when Q has the following form: Q = {λ R n : λ R}. Theorem 3.5: [6] Assume that Λ is nonempty and bounded. Then, the sequence {λ k } k 0 generated by Algorithm 3.2 is bounded. Since Assumption 2.3 holds, then Λ is nonempty. Conditions under which Λ is bounded can be found in [10] (page 301). Under the assumptions of Theorem 3.5, it follows that there exists R > 0 sufficiently large such that the set Q = {λ R n : λ R} contains Λ, and thus we can assume D Q to be finite. Notice that similar arguments were used in order to prove convergence of two-level algorithms for convex optimization problems (see e.g. [3]). ) Denote α = 4 D Q (D X + D Z )( A 2 σ Qσ X σ Qσ Z. The next theorem shows how to choose optimally the smoothness parameter c and provides the complexity estimates of our method when Q is a ball. Our proof follows a similar reasoning as in [8], but requires more complex arguments. Theorem 3.6: Assume that there exists R such that the set Q = {λ R n : λ R} contains a λ Λ. Taking c = 2 DQ ( A 2 ) k+1 D X+D Z σ Qσ X σ Qσ Z, then after k iterations we obtain an approximate solution to the problem (1) (ˆx,ẑ) = k 2(l+1) (k+1)(k+2) (xl+1,z l+1 ) and ˆλ = λ k which satisfy the following bounds on the duality gap and constraints: 0 [φ 1 (ˆx) + φ 2 (ẑ)] f 0 (ˆλ) α k + 1 α Aˆx + Bẑ b (R λ )(k + 1). Proof: Using (9) and the form of Q we obtain that max λ Q 4L c σ Q (k + 1) 2 d Q(λ) + Aˆx + Bẑ b,λ 4L cd Q σ Q (k + 1) 2 + max Aˆx + Bẑ b,λ = λ R 4L cd Q σ Q (k + 1) 2 + R Aˆx + Bẑ b. In view of the previous relation and Theorem 3.4 we obtain the following duality gap: [φ 1 (ˆx) + φ 2 (ẑ)] f 0 (ˆλ) c(d X + D Z ) + 4L cd Q σ Q (k + 1) 2 R Aˆx + Bẑ b c(d X + D Z ) + 4L cd Q σ Q (k + 1) 2 Minimizing the right-hand side of this inequality over c we get the above expressions for c and for the upper bound on the duality gap from the theorem. Moreover, for the constraints using Lemma 3.3 and inequality (7) we have that (R λ ) Aˆx+Bẑ b c(d X +D Z )+ 4L cd Q σ Q (k + 1) 2 and replacing c derived above we also get the bound on the constraints violation. From Theorem 3.6 we obtain that the complexity for finding an (ǫ,ǫ c )-approximation of the optimum f, when the set Q is a ball, is k + 1 = 4 D Q (D X + D Z ) ( A ) 2 σ Qσ X 1 σ Qσ Z ǫ, i.e. the efficiency estimates of our scheme is of the order O( 1 ǫ ), better than most non-smooth optimization schemes such as the subgradient method that have an efficiency estimate of the order O( 1 ǫ ) 2 (see e.g. [7]). The dependence of the parameters c and L c ǫ on ǫ is as follows: c = 2(D X+D Z) and L c = ( A 2 σ X + B ) 2 DX+D Z σ Z 2ǫ. The main advantage of our scheme is that it is fully automatic, the parameter c is chosen unambiguously, which is crucial for justifying the convergence properties of Algorithm 3.2. Another advantage of the proximal center method is that we are free in the choice of the norms in the spaces R n, R m and R p, while most of the decomposition schemes are based on the Euclidean norm. Thus, we can choose the norms which make the ratio A 2 σ Qσ X as small as possible. D. Efficiency estimates for the Euclidean norm In this section we assume the Euclidean norm on R n. 3080

5 Theorem 3.7: Assume that Q = R n and d Q (λ) = 1 2 λ 2, with the Euclidean norm on R n ǫ. Taking c = D X+D Z and ( k + 1 = 2 A 2 ) σ X σ (DX Z + D Z ) 1 ǫ, then after k iterations the duality gap is less than ǫ and the constraints satisfy Aˆx + Bẑ b ǫ ( λ + λ ). Proof: Let us note that for these choices for Q and d Q we have σ Q = 1 and thus max 4L c λ R n σ Q (k + 1) 2 d Q(λ) + Aˆx + Bẑ b,λ = (k + 1) 2 Aˆx + Bẑ b 2. We obtain the following bound on the duality gap (Th. 3.4): [φ 1 (ˆx) + φ 2 (ẑ)] f 0 (ˆλ) (k + 1)2 c(d X + D Z ) Aˆx + Bẑ b 2 c(d X + D Z ). ǫ It follows that taking c = D X+D Z, the duality gap is less than ǫ. For the constraints using Lemma 3.3 and inequality (7) we get that Aˆx + Bẑ b satisfies the second order (k+1) inequality in y: 2 y 2 λ y ǫ 0. Therefore, Aˆx + Bẑ b must be less than the largest root of the corresponding second-order equation, i.e. Aˆx + Bẑ b ( λ + λ 2 ǫ(k + 1)2 ) 4L c + 2L c (k + 1) 2. After some long but straightforward computations we get that after k iterations, where k defined as in the theorem, we also get Aˆx + Bẑ b ǫ ( λ + λ ). Remark 3.8 When coupling inequalities Ax + Bz b 0 are present, then we choose Q = R n +. Using the same reasoning as before we get that max λ 0 4Lc σ Q(k+1) d 2 Q (λ)+ Aˆx + Bẑ b,λ = (k+1)2 [Aˆx + Bẑ b] + 2, where [y] + denotes the projection of y R n onto R n +. Taking for c and k the same values as in Theorem 3.7, we can conclude that after k iterations the duality gap is less than ǫ and using a modified version of Lemma 3.3 (i.e. a generalized version of Cauchy-Schwarz inequality: y,λ [y] + λ for any y R n,λ 0) the constraints violation satisfy [Aˆx + Bẑ b] + ǫ ( λ + λ ). IV. APPLICATIONS A. Applications with separable structure In this section we briefly discuss some of the applications to which our method can be applied. First application that we will discuss here is the control of large scale systems with interacting subsystem dynamics. A distributed MPC framework is appealing in this context since this framework allows us to design local subsystembased controllers that take care of the interactions between different subsystems and physical constraints. We assume that the overall system model can be decomposed into M appropriate subsystem models: x i (k + 1) = A ij x j (k) + B ij u j (k) i = 1 M, j N(i) where N(i) denotes the set of subsystems that interact with the ith subsystem, including itself. The control and state sequence must satisfy local constraints x i (k) Ω i, u i (k) U i i = 1 M and k 0, where the sets Ω i and U i are usually convex compact sets with the origin in their interior. In general the control objective is to steer the state of the system to origin or any other set point in a best way. Performance is expressed via a stage cost, which in this paper we assume to have the following form (see also [13]): M l i(x i,u i ), where usually l i is a convex quadratic function, but not necessarily strictly convex. Let N denote the prediction horizon. In MPC we must solve at each step k, given x i (k) = x i, a finitehorizon optimal control problem of the following form: N 1 min l i (x i l,u i l) (10) x i l,ui l s.t. : x i 0 = x i, x i l+1 = A ij x j l + B iju j l j N(i) x i N Ω i, x i l Ω i, u i l U i l = 0 N 1, i = 1 M Note that a similar formulation of distributed MPC for coupled linear subsystems with decoupled costs was given in [13], but without state constraints (i.e. without imposing x i (k) Ω i ). Let us introduce x i = (x i 0 x i N ui 0 u i N 1 ),X i = Ω N+1 i Ui N and ψ i (x i ) = N 1 l i(x i l,ui l ). Then, the control problem (10) can be recast as a separable convex program: { M min ψ i (x i ) : C i x i γ = 0 }, (11) x i X i where the matrices C i s are defined appropriately. Network optimization furnishes another areas in which our Algorithm 3.2 leads to a new method of solution. In this application the convex optimization problem has the following form [9], [14]: { M min ψ i (x i ) : C i x i γ = 0, x i X i D i x i β 0 }, (12) where X i are compact convex sets (in general balls) in R m, ψ i s are non-strictly convex functions and M denotes the number of agents in the network. Note that (11) is a particular case of the separable convex problem (12). In [13] the optimization problem (10) (or equivalently (11)) was solved in a decentralized fashion, iterating the Jacobi algorithm p max times [1]. But, there is no theoretical guarantee of the Jacobi algorithm about how good is the approximation of the optimum after p max iterations and moreover we need strictly convex functions ψ i to prove asymptotic convergence to the optimum. However, if we solve (10) using our Algorithm 3.2, we have a guaranteed upper bound on the approximation after p max iterations (see Theorem 3.4). In [9], [14] the optimization problem (12) is 3081

6 solved using the dual subgradient method described in the Algorithm 2.2. From the simulation tests of Section IV-B we see that our Algorithm 3.2 is superior to Algorithm 2.2. Let us describe briefly the main ingredients of Algorithm 3.2 for problem (12). Let d Xi be properly chosen proxfunctions, according to the structure of the sets X i s and the norm 1 used on R m. In this case the smooth dual function f c has the form: f c (λ) = min ψ i (x i ) + λ 1, C i x i α + x i X i λ 2, D i x i β + c d Xi (x i ). Moreover, Q (R n1 R n2 + ), where n 1 and n 2 denote the number of equality and inequality constraints. It is worth noting that in this case all the minimization problems involved in Algorithm 3.2 are decomposable in x i and thus the agents can solve the optimization problem (12) in a decentralized fashion. B. Computational results We conclude this paper with the results of computational experiments on a random set of problems of the form (12), where ψ i s are convex quadratic functions (but not strictly convex) and X i s are balls in R m defined with the Euclidean norm. Similarly, R n (corresponding to the Lagrange multipliers) is endowed with the Euclidean norm. m nr iter Alg 3.2 nr iter Alg (0.01) 5000(0.31) (0.01) 5000(0.49) (0.01) 10000(0.57) (0.0011) 5000(0.31) (0.004) 5000(0.49) (0.007) 10000(0.57) In the table we display the number of iterations of the Algorithm 2.2 and 3.2 for different values of m = 50,100 and 1000, and for M = 10 agents. For m = 50 and m = 100 the maximum number of iterations that we allow is For m = 1000 we iterate at most times. We also display between brackets the corresponding accuracy. We see that the accuracy of the approximation of the optimum is much better with our Algorithm 3.2 than with Algorithm 2.2. V. CONCLUSIONS A new decomposition method in convex programming is developed in this paper using the framework of dual decomposition. Our method combines the computationally non-expensive cost of the gradient method with the efficiency of structural optimization for solving non-smooth separable convex programs. Although our decomposition method 1 For example if X i = {x R m : x x 0 2 R}, where denotes here the Euclidean norm, then it is natural to take d Xi (x) = x x If X i = {x R m P : x 0, m P x (i) = 1}, then the norm x = m x (i) and d Xi (x) = ln m + P m x (i) ln x (i) is more suitable (see [8] for more details). resembles proximal based methods, it differs both in the computational steps and in the choice of the parameters of the method. Contrary to most proximal based methods that enforce the next iterates to be close to the previous ones, our method uses fixed centers in the prox-terms which leads to more freedom in the next iterates. Another advantage of our proximal center method is that it is fully automatic, i.e. the parameters of the scheme are chosen optimally, which are crucial for justifying the convergence properties of the new scheme. We presented efficiency estimate results of the new decomposition method for general separable convex problems. We also show that our algorithm can be used to solve network optimization or distributed model predictive control problems. The computations on some test problems confirm that the proposed method works well in practice. Acknowledgment. We acknowledge the financial support by Research Council K.U. Leuven: GOA AMBioRICS, CoE EF/05/006, OT/03/12, PhD/postdoc & fellow grants; Flemish Government: FWO PhD/postdoc grants, FWO projects G , G , G , G ; Research communities (ICCoS, ANMMM, MLDM); AWI: BIL/05/43, IWT: PhD Grants; Belgian Federal Science Policy Office: IUAP DYSCO. REFERENCES [1] D. P. Bertsekas and J. N. Tsitsiklis. Parallel and distributed computation: Numerical Methods. Prentice-Hall, Englewood Cliffs, NJ, [2] G. Chen and M. Teboulle. A proximal-based decomposition method for convex minimization problems. Mathematical Programming (A), 64:81 101, [3] G. Cohen. Optimization by decomposition and coordination: A unified approach. IEEE Transactions on Automatic Control, AC 23(2): , [4] J. Eckstein and D. Bertsekas. On the Douglas Rachford splitting method and the proximal point method for maximal monotone operators. Mathematical Programming (A), 55(3): , [5] S. Kontogiorgis, R. De Leone, and R. Meyer. Alternating direction splitings for block angular parallel optimization. Journal of Optimization Theory and Applications, 90(1):1 29, [6] I. Necoara and J.A.K. Suykens. Application of a smoothing technique to decomposition in convex programming. Technical Report 08 07, ESAT-SISTA, K.U. Leuven (Leuven, Belgium), January [7] Y. Nesterov. Introductory Lectures on Convex Optimization: A Basic Course. Kluwer, Boston, [8] Y. Nesterov. Smooth minimization of non-smooth functions. Mathematical Programming (A), 103(1): , [9] D. Palomar and M. Chiang. Alternative distributed algorithms for network utility maximization: Framework and applications. IEEE Transactions on Automatic Control, 52(12): , [10] R. T. Rockafellar. Convex Analysis, volume 28 of Princeton Mathematics Series. Princeton University Press, [11] J. E. Spingarn. Applications of the method of partial inverses to convex programming: decomposition. Mathematical Programming (A), 32: , [12] P. Tseng. Applications of a splitting algorithm to decomposition in convex programming and variational inequalities. SIAM Journal of Control and Optimization, 29(1): , [13] A. Venkat, I. Hiskens, J. Rawlings, and S. Wright. Distributed MPC strategies with application to power system automatic generation control. IEEE Transactions on Control Systems Technology, to appear, [14] L. Xiao, M. Johansson, and S. Boyd. Simultaneous routing and resource allocation via dual decomposition. IEEE Transactions on Communications, 52(7): ,

AN IMPROVED DUAL DECOMPOSITION APPROACH TO DSL DYNAMIC SPECTRUM MANAGEMENT

AN IMPROVED DUAL DECOMPOSITION APPROACH TO DSL DYNAMIC SPECTRUM MANAGEMENT AN IMPROVED DUAL DECOMPOSITION APPROACH TO DSL DYNAMIC SPECTRUM MANAGEMENT Paschalis Tsiaflais, Ion Necoara, Johan A. K. Suyens, and Marc Moonen Katholiee Universiteit Leuven, Electrical Engineering, ESAT-SCD

More information

Lecture 4 Duality and Decomposition Techniques

Lecture 4 Duality and Decomposition Techniques Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian

More information

Optimization for Machine Learning

Optimization for Machine Learning Optimization for Machine Learning (Problems; Algorithms - C) SUVRIT SRA Massachusetts Institute of Technology PKU Summer School on Data Science (July 2017) Course materials http://suvrit.de/teaching.html

More information

Convex optimization algorithms for sparse and low-rank representations

Convex optimization algorithms for sparse and low-rank representations Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,

More information

Lagrangian Relaxation: An overview

Lagrangian Relaxation: An overview Discrete Math for Bioinformatics WS 11/12:, by A. Bockmayr/K. Reinert, 22. Januar 2013, 13:27 4001 Lagrangian Relaxation: An overview Sources for this lecture: D. Bertsimas and J. Tsitsiklis: Introduction

More information

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics 1. Introduction EE 546, Univ of Washington, Spring 2016 performance of numerical methods complexity bounds structural convex optimization course goals and topics 1 1 Some course info Welcome to EE 546!

More information

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal

More information

Distributed Alternating Direction Method of Multipliers

Distributed Alternating Direction Method of Multipliers Distributed Alternating Direction Method of Multipliers Ermin Wei and Asuman Ozdaglar Abstract We consider a network of agents that are cooperatively solving a global unconstrained optimization problem,

More information

A primal-dual framework for mixtures of regularizers

A primal-dual framework for mixtures of regularizers A primal-dual framework for mixtures of regularizers Baran Gözcü baran.goezcue@epfl.ch Laboratory for Information and Inference Systems (LIONS) École Polytechnique Fédérale de Lausanne (EPFL) Switzerland

More information

Convexity Theory and Gradient Methods

Convexity Theory and Gradient Methods Convexity Theory and Gradient Methods Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Convex Functions Optimality

More information

Mathematical Programming and Research Methods (Part II)

Mathematical Programming and Research Methods (Part II) Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types

More information

Convex Optimization MLSS 2015

Convex Optimization MLSS 2015 Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :

More information

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.

More information

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007 : Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers With Adaptive Step Size Selection Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein October 8, 2015 1 / 30 Introduction Presentation Outline 1 Convex

More information

Convex Optimization and Machine Learning

Convex Optimization and Machine Learning Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i

More information

Math 5593 Linear Programming Lecture Notes

Math 5593 Linear Programming Lecture Notes Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................

More information

The Alternating Direction Method of Multipliers

The Alternating Direction Method of Multipliers The Alternating Direction Method of Multipliers Customizable software solver package Peter Sutor, Jr. Project Advisor: Professor Tom Goldstein April 27, 2016 1 / 28 Background The Dual Problem Consider

More information

Lecture 2 September 3

Lecture 2 September 3 EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give

More information

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A.

ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS. Donghwan Kim and Jeffrey A. ACCELERATED DUAL GRADIENT-BASED METHODS FOR TOTAL VARIATION IMAGE DENOISING/DEBLURRING PROBLEMS Donghwan Kim and Jeffrey A. Fessler University of Michigan Dept. of Electrical Engineering and Computer Science

More information

Characterizing Improving Directions Unconstrained Optimization

Characterizing Improving Directions Unconstrained Optimization Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,

More information

Introduction to Modern Control Systems

Introduction to Modern Control Systems Introduction to Modern Control Systems Convex Optimization, Duality and Linear Matrix Inequalities Kostas Margellos University of Oxford AIMS CDT 2016-17 Introduction to Modern Control Systems November

More information

Convexity: an introduction

Convexity: an introduction Convexity: an introduction Geir Dahl CMA, Dept. of Mathematics and Dept. of Informatics University of Oslo 1 / 74 1. Introduction 1. Introduction what is convexity where does it arise main concepts and

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

Gate Sizing by Lagrangian Relaxation Revisited

Gate Sizing by Lagrangian Relaxation Revisited Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

LECTURE 18 LECTURE OUTLINE

LECTURE 18 LECTURE OUTLINE LECTURE 18 LECTURE OUTLINE Generalized polyhedral approximation methods Combined cutting plane and simplicial decomposition methods Lecture based on the paper D. P. Bertsekas and H. Yu, A Unifying Polyhedral

More information

PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY

PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY PROJECTION ONTO A POLYHEDRON THAT EXPLOITS SPARSITY WILLIAM W. HAGER AND HONGCHAO ZHANG Abstract. An algorithm is developed for projecting a point onto a polyhedron. The algorithm solves a dual version

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Mathematical and Algorithmic Foundations Linear Programming and Matchings Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis

More information

AM 221: Advanced Optimization Spring 2016

AM 221: Advanced Optimization Spring 2016 AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 2 Wednesday, January 27th 1 Overview In our previous lecture we discussed several applications of optimization, introduced basic terminology,

More information

Improved Dual Decomposition Based. Optimization for DSL Dynamic Spectrum Management

Improved Dual Decomposition Based. Optimization for DSL Dynamic Spectrum Management Improved Dual Decomposition Based 1 Optimization for DSL Dynamic Spectrum Management arxiv:1302.3146v1 [math.oc] 13 Feb 2013 Paschalis Tsiaflais, Ion Necoara, Johan A. K. Suyens, Marc Moonen Electrical

More information

Convex Optimization. Lijun Zhang Modification of

Convex Optimization. Lijun Zhang   Modification of Convex Optimization Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Modification of http://stanford.edu/~boyd/cvxbook/bv_cvxslides.pdf Outline Introduction Convex Sets & Functions Convex Optimization

More information

Proximal operator and methods

Proximal operator and methods Proximal operator and methods Master 2 Data Science, Univ. Paris Saclay Robert M. Gower Optimization Sum of Terms A Datum Function Finite Sum Training Problem The Training Problem Convergence GD I Theorem

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

Lecture 2. Topology of Sets in R n. August 27, 2008

Lecture 2. Topology of Sets in R n. August 27, 2008 Lecture 2 Topology of Sets in R n August 27, 2008 Outline Vectors, Matrices, Norms, Convergence Open and Closed Sets Special Sets: Subspace, Affine Set, Cone, Convex Set Special Convex Sets: Hyperplane,

More information

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form Philip E. Gill Vyacheslav Kungurtsev Daniel P. Robinson UCSD Center for Computational Mathematics Technical Report

More information

Research Interests Optimization:

Research Interests Optimization: Mitchell: Research interests 1 Research Interests Optimization: looking for the best solution from among a number of candidates. Prototypical optimization problem: min f(x) subject to g(x) 0 x X IR n Here,

More information

Lecture 18: March 23

Lecture 18: March 23 0-725/36-725: Convex Optimization Spring 205 Lecturer: Ryan Tibshirani Lecture 8: March 23 Scribes: James Duyck Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

Numerical schemes for Hamilton-Jacobi equations, control problems and games

Numerical schemes for Hamilton-Jacobi equations, control problems and games Numerical schemes for Hamilton-Jacobi equations, control problems and games M. Falcone H. Zidani SADCO Spring School Applied and Numerical Optimal Control April 23-27, 2012, Paris Lecture 2/3 M. Falcone

More information

ME 555: Distributed Optimization

ME 555: Distributed Optimization ME 555: Distributed Optimization Duke University Spring 2015 1 Administrative Course: ME 555: Distributed Optimization (Spring 2015) Instructor: Time: Location: Office hours: Website: Soomin Lee (email:

More information

Global Minimization via Piecewise-Linear Underestimation

Global Minimization via Piecewise-Linear Underestimation Journal of Global Optimization,, 1 9 (2004) c 2004 Kluwer Academic Publishers, Boston. Manufactured in The Netherlands. Global Minimization via Piecewise-Linear Underestimation O. L. MANGASARIAN olvi@cs.wisc.edu

More information

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material

More information

AVERAGING RANDOM PROJECTION: A FAST ONLINE SOLUTION FOR LARGE-SCALE CONSTRAINED STOCHASTIC OPTIMIZATION. Jialin Liu, Yuantao Gu, and Mengdi Wang

AVERAGING RANDOM PROJECTION: A FAST ONLINE SOLUTION FOR LARGE-SCALE CONSTRAINED STOCHASTIC OPTIMIZATION. Jialin Liu, Yuantao Gu, and Mengdi Wang AVERAGING RANDOM PROJECTION: A FAST ONLINE SOLUTION FOR LARGE-SCALE CONSTRAINED STOCHASTIC OPTIMIZATION Jialin Liu, Yuantao Gu, and Mengdi Wang Tsinghua National Laboratory for Information Science and

More information

Convergence of Nonconvex Douglas-Rachford Splitting and Nonconvex ADMM

Convergence of Nonconvex Douglas-Rachford Splitting and Nonconvex ADMM Convergence of Nonconvex Douglas-Rachford Splitting and Nonconvex ADMM Shuvomoy Das Gupta Thales Canada 56th Allerton Conference on Communication, Control and Computing, 2018 1 What is this talk about?

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 1 A. d Aspremont. Convex Optimization M2. 1/49 Today Convex optimization: introduction Course organization and other gory details... Convex sets, basic definitions. A. d

More information

Learning with infinitely many features

Learning with infinitely many features Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example

More information

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction

More information

Fast-Lipschitz Optimization

Fast-Lipschitz Optimization Fast-Lipschitz Optimization DREAM Seminar Series University of California at Berkeley September 11, 2012 Carlo Fischione ACCESS Linnaeus Center, Electrical Engineering KTH Royal Institute of Technology

More information

60 2 Convex sets. {x a T x b} {x ã T x b}

60 2 Convex sets. {x a T x b} {x ã T x b} 60 2 Convex sets Exercises Definition of convexity 21 Let C R n be a convex set, with x 1,, x k C, and let θ 1,, θ k R satisfy θ i 0, θ 1 + + θ k = 1 Show that θ 1x 1 + + θ k x k C (The definition of convexity

More information

Lecture 2 - Introduction to Polytopes

Lecture 2 - Introduction to Polytopes Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.

More information

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous

More information

Lecture 19 Subgradient Methods. November 5, 2008

Lecture 19 Subgradient Methods. November 5, 2008 Subgradient Methods November 5, 2008 Outline Lecture 19 Subgradients and Level Sets Subgradient Method Convergence and Convergence Rate Convex Optimization 1 Subgradients and Level Sets A vector s is a

More information

Solution Methods Numerical Algorithms

Solution Methods Numerical Algorithms Solution Methods Numerical Algorithms Evelien van der Hurk DTU Managment Engineering Class Exercises From Last Time 2 DTU Management Engineering 42111: Static and Dynamic Optimization (6) 09/10/2017 Class

More information

Introduction to Constrained Optimization

Introduction to Constrained Optimization Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR

More information

A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material

A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material A MRF Shape Prior for Facade Parsing with Occlusions Supplementary Material Mateusz Koziński, Raghudeep Gadde, Sergey Zagoruyko, Guillaume Obozinski and Renaud Marlet Université Paris-Est, LIGM (UMR CNRS

More information

Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization. Author: Martin Jaggi Presenter: Zhongxing Peng

Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization. Author: Martin Jaggi Presenter: Zhongxing Peng Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization Author: Martin Jaggi Presenter: Zhongxing Peng Outline 1. Theoretical Results 2. Applications Outline 1. Theoretical Results 2. Applications

More information

A Tutorial on Decomposition Methods for Network Utility Maximization Daniel P. Palomar, Member, IEEE, and Mung Chiang, Member, IEEE.

A Tutorial on Decomposition Methods for Network Utility Maximization Daniel P. Palomar, Member, IEEE, and Mung Chiang, Member, IEEE. IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, VOL. 24, NO. 8, AUGUST 2006 1439 A Tutorial on Decomposition Methods for Network Utility Maximization Daniel P. Palomar, Member, IEEE, and Mung Chiang,

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

On Efficient Optimal Transport: An Analysis of Greedy and Accelerated Mirror Descent Algorithms

On Efficient Optimal Transport: An Analysis of Greedy and Accelerated Mirror Descent Algorithms On Efficient Optimal Transport: An Analysis of Greedy and Accelerated Mirror Descent Algorithms Tianyi Lin, Nhat Ho, Michael I Jordan, Department of Electrical Engineering and Computer Sciences Department

More information

Alternating Projections

Alternating Projections Alternating Projections Stephen Boyd and Jon Dattorro EE392o, Stanford University Autumn, 2003 1 Alternating projection algorithm Alternating projections is a very simple algorithm for computing a point

More information

Alternating Direction Method of Multipliers

Alternating Direction Method of Multipliers Alternating Direction Method of Multipliers CS 584: Big Data Analytics Material adapted from Stephen Boyd (https://web.stanford.edu/~boyd/papers/pdf/admm_slides.pdf) & Ryan Tibshirani (http://stat.cmu.edu/~ryantibs/convexopt/lectures/21-dual-meth.pdf)

More information

Machine Learning for Signal Processing Lecture 4: Optimization

Machine Learning for Signal Processing Lecture 4: Optimization Machine Learning for Signal Processing Lecture 4: Optimization 13 Sep 2015 Instructor: Bhiksha Raj (slides largely by Najim Dehak, JHU) 11-755/18-797 1 Index 1. The problem of optimization 2. Direct optimization

More information

Lecture 19: November 5

Lecture 19: November 5 0-725/36-725: Convex Optimization Fall 205 Lecturer: Ryan Tibshirani Lecture 9: November 5 Scribes: Hyun Ah Song Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling

DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS. Yujiao Cheng, Houfeng Huang, Gang Wu, Qing Ling DISTRIBUTED NETWORK RESOURCE ALLOCATION WITH INTEGER CONSTRAINTS Yuao Cheng, Houfeng Huang, Gang Wu, Qing Ling Department of Automation, University of Science and Technology of China, Hefei, China ABSTRACT

More information

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs 15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest

More information

Parallel Auction Algorithm for Linear Assignment Problem

Parallel Auction Algorithm for Linear Assignment Problem Parallel Auction Algorithm for Linear Assignment Problem Xin Jin 1 Introduction The (linear) assignment problem is one of classic combinatorial optimization problems, first appearing in the studies on

More information

Linear Bilevel Programming With Upper Level Constraints Depending on the Lower Level Solution

Linear Bilevel Programming With Upper Level Constraints Depending on the Lower Level Solution Linear Bilevel Programming With Upper Level Constraints Depending on the Lower Level Solution Ayalew Getachew Mersha and Stephan Dempe October 17, 2005 Abstract Focus in the paper is on the definition

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

Optimal network flow allocation

Optimal network flow allocation Optimal network flow allocation EE384Y Project intermediate report Almir Mutapcic and Primoz Skraba Stanford University, Spring 2003-04 May 10, 2004 Contents 1 Introduction 2 2 Background 2 3 Problem statement

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

Tutorial on Convex Optimization for Engineers

Tutorial on Convex Optimization for Engineers Tutorial on Convex Optimization for Engineers M.Sc. Jens Steinwandt Communications Research Laboratory Ilmenau University of Technology PO Box 100565 D-98684 Ilmenau, Germany jens.steinwandt@tu-ilmenau.de

More information

A More Efficient Approach to Large Scale Matrix Completion Problems

A More Efficient Approach to Large Scale Matrix Completion Problems A More Efficient Approach to Large Scale Matrix Completion Problems Matthew Olson August 25, 2014 Abstract This paper investigates a scalable optimization procedure to the low-rank matrix completion problem

More information

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming Gianni Di Pillo (dipillo@dis.uniroma1.it) Giampaolo Liuzzi (liuzzi@iasi.cnr.it) Stefano Lucidi (lucidi@dis.uniroma1.it)

More information

An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation

An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation Dhaka Univ. J. Sci. 61(2): 135-140, 2013 (July) An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation M. Babul Hasan and Md. Toha De epartment of Mathematics, Dhaka

More information

Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology

Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology Aspects of Convex, Nonconvex, and Geometric Optimization (Lecture 1) Suvrit Sra Massachusetts Institute of Technology Hausdorff Institute for Mathematics (HIM) Trimester: Mathematics of Signal Processing

More information

11 Linear Programming

11 Linear Programming 11 Linear Programming 11.1 Definition and Importance The final topic in this course is Linear Programming. We say that a problem is an instance of linear programming when it can be effectively expressed

More information

Probabilistic Graphical Models

Probabilistic Graphical Models School of Computer Science Probabilistic Graphical Models Theory of Variational Inference: Inner and Outer Approximation Eric Xing Lecture 14, February 29, 2016 Reading: W & J Book Chapters Eric Xing @

More information

Projection-Based Methods in Optimization

Projection-Based Methods in Optimization Projection-Based Methods in Optimization Charles Byrne (Charles Byrne@uml.edu) http://faculty.uml.edu/cbyrne/cbyrne.html Department of Mathematical Sciences University of Massachusetts Lowell Lowell, MA

More information

ONLY AVAILABLE IN ELECTRONIC FORM

ONLY AVAILABLE IN ELECTRONIC FORM MANAGEMENT SCIENCE doi 10.1287/mnsc.1070.0812ec pp. ec1 ec7 e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 2008 INFORMS Electronic Companion Customized Bundle Pricing for Information Goods: A Nonlinear

More information

Iteratively Re-weighted Least Squares for Sums of Convex Functions

Iteratively Re-weighted Least Squares for Sums of Convex Functions Iteratively Re-weighted Least Squares for Sums of Convex Functions James Burke University of Washington Jiashan Wang LinkedIn Frank Curtis Lehigh University Hao Wang Shanghai Tech University Daiwei He

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers

Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Stephen Boyd MIIS Xi An, 1/7/12 source: Distributed Optimization and Statistical Learning via the Alternating

More information

OPERATIONS RESEARCH. Linear Programming Problem

OPERATIONS RESEARCH. Linear Programming Problem OPERATIONS RESEARCH Chapter 1 Linear Programming Problem Prof. Bibhas C. Giri Department of Mathematics Jadavpur University Kolkata, India Email: bcgiri.jumath@gmail.com 1.0 Introduction Linear programming

More information

Convexity. 1 X i is convex. = b is a hyperplane in R n, and is denoted H(p, b) i.e.,

Convexity. 1 X i is convex. = b is a hyperplane in R n, and is denoted H(p, b) i.e., Convexity We ll assume throughout, without always saying so, that we re in the finite-dimensional Euclidean vector space R n, although sometimes, for statements that hold in any vector space, we ll say

More information

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Parallelization in Non-hierarchical Analytical Target Cascading Yongsu Jung Department of Mechanical Engineering,

More information

Convex Analysis and Minimization Algorithms I

Convex Analysis and Minimization Algorithms I Jean-Baptiste Hiriart-Urruty Claude Lemarechal Convex Analysis and Minimization Algorithms I Fundamentals With 113 Figures Springer-Verlag Berlin Heidelberg New York London Paris Tokyo Hong Kong Barcelona

More information

Convex Optimization / Homework 2, due Oct 3

Convex Optimization / Homework 2, due Oct 3 Convex Optimization 0-725/36-725 Homework 2, due Oct 3 Instructions: You must complete Problems 3 and either Problem 4 or Problem 5 (your choice between the two) When you submit the homework, upload a

More information

A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems

A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems Keely L. Croxton Fisher College of Business The Ohio State University Bernard Gendron Département

More information

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem Computational Learning Theory Fall Semester, 2012/13 Lecture 10: SVM Lecturer: Yishay Mansour Scribe: Gitit Kehat, Yogev Vaknin and Ezra Levin 1 10.1 Lecture Overview In this lecture we present in detail

More information

Lecture 5: Duality Theory

Lecture 5: Duality Theory Lecture 5: Duality Theory Rajat Mittal IIT Kanpur The objective of this lecture note will be to learn duality theory of linear programming. We are planning to answer following questions. What are hyperplane

More information

Algorithms for convex optimization

Algorithms for convex optimization Algorithms for convex optimization Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University kocvara@utia.cas.cz http://www.utia.cas.cz/kocvara

More information

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms

Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms Iterative Algorithms I: Elementary Iterative Methods and the Conjugate Gradient Algorithms By:- Nitin Kamra Indian Institute of Technology, Delhi Advisor:- Prof. Ulrich Reude 1. Introduction to Linear

More information

3 Interior Point Method

3 Interior Point Method 3 Interior Point Method Linear programming (LP) is one of the most useful mathematical techniques. Recent advances in computer technology and algorithms have improved computational speed by several orders

More information

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.

More information

DISTRIBUTED ALGORITHMS FOR BASIS PURSUIT. Institute for Systems and Robotics / IST, Lisboa, Portugal

DISTRIBUTED ALGORITHMS FOR BASIS PURSUIT. Institute for Systems and Robotics / IST, Lisboa, Portugal DISTRIBUTED ALGORITHMS FOR BASIS PURSUIT João F C Mota, João M F Xavier, Pedro M Q Aguiar, Markus Püschel Institute for Systems and Robotics / IST, Lisboa, Portugal {jmota,jxavier,aguiar}@isristutlpt Department

More information

Dual Subgradient Methods Using Approximate Multipliers

Dual Subgradient Methods Using Approximate Multipliers Dual Subgradient Methods Using Approximate Multipliers Víctor Valls, Douglas J. Leith Trinity College Dublin Abstract We consider the subgradient method for the dual problem in convex optimisation with

More information