Introduction to Constrained Optimization

Similar documents
Applied Lagrange Duality for Constrained Optimization

Introduction to Machine Learning

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Kernels and Constrained Optimization

Convex Optimization and Machine Learning

Convex Optimization. Lijun Zhang Modification of

Optimization III: Constrained Optimization

Linear methods for supervised learning

Programming, numerics and optimization

Lecture 7: Support Vector Machine

COMS 4771 Support Vector Machines. Nakul Verma

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Mathematical Programming and Research Methods (Part II)

Lecture 15: Log Barrier Method

Characterizing Improving Directions Unconstrained Optimization

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

Demo 1: KKT conditions with inequality constraints

Constrained Optimization

Optimization under uncertainty: modeling and solution methods

Nonlinear Programming

California Institute of Technology Crash-Course on Convex Optimization Fall Ec 133 Guilherme Freitas

Lecture 5: Duality Theory

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma

CME307/MS&E311 Theory Summary

Lec 11 Rate-Distortion Optimization (RDO) in Video Coding-I

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Convex Optimization. Erick Delage, and Ashutosh Saxena. October 20, (a) (b) (c)

Introduction to Modern Control Systems

A Short SVM (Support Vector Machine) Tutorial

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Introduction to Optimization

Convexity Theory and Gradient Methods

ISM206 Lecture, April 26, 2005 Optimization of Nonlinear Objectives, with Non-Linear Constraints

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem

Linear Programming. Larry Blume. Cornell University & The Santa Fe Institute & IHS

Solution Methods Numerical Algorithms

College of Computer & Information Science Fall 2007 Northeastern University 14 September 2007

Lecture 4 Duality and Decomposition Techniques

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

In other words, we want to find the domain points that yield the maximum or minimum values (extrema) of the function.

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs

Introduction to Optimization

CME307/MS&E311 Optimization Theory Summary

9. Support Vector Machines. The linearly separable case: hard-margin SVMs. The linearly separable case: hard-margin SVMs. Learning objectives

QEM Optimization, WS 2017/18 Part 4. Constrained optimization

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Programs. Introduction

Constrained optimization

5 Day 5: Maxima and minima for n variables.

Convex Optimization MLSS 2015

Lecture Notes: Constraint Optimization

Lecture 2 Optimization with equality constraints

Linear programming and duality theory

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

11 Linear Programming

Gate Sizing by Lagrangian Relaxation Revisited

Section Notes 5. Review of Linear Programming. Applied Math / Engineering Sciences 121. Week of October 15, 2017

Lagrangian Relaxation: An overview

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21

Projection onto the probability simplex: An efficient algorithm with a simple proof, and an application

Kernel Methods & Support Vector Machines

Math 5593 Linear Programming Lecture Notes

Introduction to Mathematical Programming IE406. Lecture 20. Dr. Ted Ralphs

Section Notes 4. Duality, Sensitivity, and the Dual Simplex Algorithm. Applied Math / Engineering Sciences 121. Week of October 8, 2018

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

Interior Point I. Lab 21. Introduction

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini

IE 521 Convex Optimization

Duality. Primal program P: Maximize n. Dual program D: Minimize m. j=1 c jx j subject to n. j=1. i=1 b iy i subject to m. i=1

A primal-dual framework for mixtures of regularizers

Repetition: Primal Dual for Set Cover

Chapter II. Linear Programming

Lagrangian Relaxation

Lecture 3: Convex sets

Elements of Economic Analysis II Lecture III: Cost Minimization, Factor Demand and Cost Function

Linear Programming Duality and Algorithms

Some Advanced Topics in Linear Programming

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

George B. Dantzig Mukund N. Thapa. Linear Programming. 1: Introduction. With 87 Illustrations. Springer

Approximation Algorithms: The Primal-Dual Method. My T. Thai

1. Lecture notes on bipartite matching February 4th,

EC5555 Economics Masters Refresher Course in Mathematics September Lecture 6 Optimization with equality constraints Francesco Feri

AM 221: Advanced Optimization Spring 2016

ORIE 6300 Mathematical Programming I September 2, Lecture 3

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Lagrangian Multipliers

Mathematical and Algorithmic Foundations Linear Programming and Matchings

CS 473: Algorithms. Ruta Mehta. Spring University of Illinois, Urbana-Champaign. Ruta (UIUC) CS473 1 Spring / 36

Introduction to optimization

Approximation Algorithms

Linear Programming. Course review MS-E2140. v. 1.1

Lecture 2 September 3

Lecture 2 - Introduction to Polytopes

Support Vector Machines.

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection

Linear Programming. Linear programming provides methods for allocating limited resources among competing activities in an optimal way.

Detecting Infeasibility in Infeasible-Interior-Point. Methods for Optimization

Linear Programming Problems

Transcription:

Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR Convex Optimization 1/14

Geometry of the Problem LNMIIT MLPR Convex Optimization 2/14

Geometry of the Problem Let us try to solve optimization problem min f (x) = min x 2 x R2 x R 2 1 + x2 2 (1) LNMIIT MLPR Convex Optimization 2/14

Geometry of the Problem Let us try to solve optimization problem subject to min f (x) = min x 2 x R2 x R 2 1 + x2 2 (1) g(x) = 2 x 1 x 2 0 (2) LNMIIT MLPR Convex Optimization 2/14

Geometry of the Problem Let us try to solve optimization problem subject to min f (x) = min x 2 x R2 x R 2 1 + x2 2 (1) g(x) = 2 x 1 x 2 0 (2) Constrained optimization demands more than stationarity at optimal point, i.e. x f (x) = 0 is not sufficient. Why? LNMIIT MLPR Convex Optimization 2/14

How do we treat Constraint? LNMIIT MLPR Convex Optimization 3/14

How do we treat Constraint? Solve this constrained optimization with a Lagrange multiplier L(x, α) = f (x) + αg(x) (3) = (x 2 1 + x 2 2 ) + α(2 x 1 x 2 ) LNMIIT MLPR Convex Optimization 3/14

How do we treat Constraint? Solve this constrained optimization with a Lagrange multiplier and the solution is L(x, α) = f (x) + αg(x) (3) = (x 2 1 + x 2 2 ) + α(2 x 1 x 2 ) x = (x 1, x 2 ) = (1, 1) (4) Before proceeding further let us first fix some notations and concepts LNMIIT MLPR Convex Optimization 3/14

Convex Optimization [1] Study of mathematical optimization problems of the form, minimize f (x) x R n subject to x C (5) x R n is a vector known as the optimization variable, f : R n R is a convex function that we want to minimize, and C R n is a convex set describing the set of feasible solutions Recall the definition of a Convex set and a Convex function 1 1 A function f : S R is convex if S is a convex set, and for any x, y S and λ [0, 1], we have f (λx + (1 λ)y) λf (x) + (1 λ)f (y). A function f is concave if f is convex. LNMIIT MLPR Convex Optimization 4/14

Lagrange Duality LNMIIT MLPR Convex Optimization 5/14

Lagrange Duality The theory of Lagrange duality is the study of optimal solutions to convex optimization problems LNMIIT MLPR Convex Optimization 5/14

Lagrange Duality The theory of Lagrange duality is the study of optimal solutions to convex optimization problems In this lecture we will discuss about differentiable convex optimization problems of the form minimize f (x) x R n subject to g i (x) 0, i = 1,..., m h i (x) = 0, i = 1,..., p (6) where f and g i are differentiable convex functions and h i are affine functions LNMIIT MLPR Convex Optimization 5/14

The Lagrangian LNMIIT MLPR Convex Optimization 6/14

The Lagrangian Intuitively, the Lagrangian can be thought of as a modified version of the original objective function. The Lagrangian L : R n R m R p R, is defined as L(x, α, β) = f (x) + m α i g i (x) + i=1 p β i h i (x). (7) i=1 LNMIIT MLPR Convex Optimization 6/14

The Lagrangian Intuitively, the Lagrangian can be thought of as a modified version of the original objective function. The Lagrangian L : R n R m R p R, is defined as L(x, α, β) = f (x) + m α i g i (x) + i=1 p β i h i (x). (7) i=1 Primal variables : x R n Dual variables : α i R m and β i R p LNMIIT MLPR Convex Optimization 6/14

The Lagrangian Intuitively, the Lagrangian can be thought of as a modified version of the original objective function. The Lagrangian L : R n R m R p R, is defined as L(x, α, β) = f (x) + m α i g i (x) + i=1 p β i h i (x). (7) i=1 Primal variables : x R n Dual variables : α i R m and β i R p The Lagrange multipliers α i and β i can be though of as costs associated with violating different constraints LNMIIT MLPR Convex Optimization 6/14

Intuition behind Lagrange Duality For any convex optimization problem, there always exist settings of the dual variables such that the unconstrained minimum of the Lagrangian with respect to the primal variables (keeping the dual variables fixed) coincides with the solution of the original constrained optimization problem [2]. LNMIIT MLPR Convex Optimization 7/14

Primal Problem Consider the optimization problem, { } min max L(x, α, β) x α,β:α i 0, i } {{ } call this θ P (x) = min x θ P (x) (8) The function θ P : R n R is called the primal objective, and the unconstrained optimization problem on the right hand side is known as the primal problem A point x R n is primal feasible if g i (x) 0, i = 1,..., m and h i (x) = 0, i = 1,..., p The optimal value of the primal objective is denoted by p = θ P (x ) and x R n is called the solution of primal problem LNMIIT MLPR Convex Optimization 8/14

Interpretation of Primal Problem 2 θ P (x) = max α,β:α i 0, i = max α,β:α i 0, i = f (x) + max α,β:α i 0, i L(x, α, β) (9) [ ] m p f (x) + α i g i (x) + β i h i (x) i=1 [ m α i g i (x) + i=1 i=1 ] p β i h i (x) i=1 2 Observe that the primal objective, θ P(x), is a convex function of x LNMIIT MLPR Convex Optimization 9/14

Interpretation of Primal Problem 2 θ P (x) = max α,β:α i 0, i = max α,β:α i 0, i = f (x) + max α,β:α i 0, i L(x, α, β) (9) [ ] m p f (x) + α i g i (x) + β i h i (x) i=1 [ m α i g i (x) + i=1 i=1 ] p β i h i (x) i=1 θ P (x) = { 0 if x is primal feasible f (x) + }{{} if x is primal infeasible original objective }{{} barrier function for carving away infeasible solutions (10) 2 Observe that the primal objective, θ P(x), is a convex function of x LNMIIT MLPR Convex Optimization 9/14

Dual Problem Switching the order of min and max we obtain an entirely different optimization problem { } max min L(x, α, β) = max θ α,β:α i 0, i } x D(α, β) (11) α,β:α {{} i 0, i call this θ D (α,β) The function θ D : R m R p R is called the dual objective, and the unconstrained optimization problem on the right hand side is known as the dual problem Generally, we say that (α, β) are dual feasible if α i 0, i = 1,..., m The optimal value of the dual objective is denoted by d = θ D (α, β ) and (α, β ) R m R p is called the solution of the dual problem LNMIIT MLPR Convex Optimization 10/14

Recall the Example Let us re-visit the example that we solved in the beginning min f (x) = min x 2 x R2 x R 2 1 + x2 2 (12) LNMIIT MLPR Convex Optimization 11/14

Recall the Example Let us re-visit the example that we solved in the beginning subject to min f (x) = min x 2 x R2 x R 2 1 + x2 2 (12) g(x) = 2 x 1 x 2 0 (13) LNMIIT MLPR Convex Optimization 11/14

Recall the Example Let us re-visit the example that we solved in the beginning subject to min f (x) = min x 2 x R2 x R 2 1 + x2 2 (12) g(x) = 2 x 1 x 2 0 (13) Do you see primal and dual formulations? LNMIIT MLPR Convex Optimization 11/14

Interpretation of Dual Problem LNMIIT MLPR Convex Optimization 12/14

Interpretation of Dual Problem Lemma 1. If (α, β) are dual feasible, then θ D (α, β) p. 2. (Weak Duality) For any pair of primal and dual problems, d p. 3. (Strong Duality) For any pair of primal and dual problems which satisfy certain technical conditions called constraint qualification, then d = p. 4. (Complementary slackness) If strong duality holds, then α i g i(x ) = 0 for each i = 1,..., m. LNMIIT MLPR Convex Optimization 12/14

The KKT Conditions [3] LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, 2. (Dual feasibility) α i 0, i = 1,..., p, LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, 2. (Dual feasibility) α i 0, i = 1,..., p, 3. (Complementary slackness) α i g i(x ) = 0 for each i = 1,..., m, and LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, 2. (Dual feasibility) α i 0, i = 1,..., p, 3. (Complementary slackness) α i g i(x ) = 0 for each i = 1,..., m, and 4. (Lagrangian stationarity) x L(x, α, β ) = 0. LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, 2. (Dual feasibility) α i 0, i = 1,..., p, 3. (Complementary slackness) α i g i(x ) = 0 for each i = 1,..., m, and 4. (Lagrangian stationarity) x L(x, α, β ) = 0. Then x is primal optimal and (α, β ) are dual optimal. furhtermore, if strong duality holds, then any primal optimal x and dual optimal (α, β ) must satisfy the conditions 1 through 4. LNMIIT MLPR Convex Optimization 13/14

The KKT Conditions [3] Theorem Suppose that x R n, α R m and β R p satisfy the following conditions: 1. (Primal feasibility) g i (x ) 0, i = 1,..., m and h i (x ) = 0, i = 1,..., p, 2. (Dual feasibility) α i 0, i = 1,..., p, 3. (Complementary slackness) α i g i(x ) = 0 for each i = 1,..., m, and 4. (Lagrangian stationarity) x L(x, α, β ) = 0. Then x is primal optimal and (α, β ) are dual optimal. furhtermore, if strong duality holds, then any primal optimal x and dual optimal (α, β ) must satisfy the conditions 1 through 4. These conditions are known as the Karush-Kuhn-Tucker (KKT) conditions. LNMIIT MLPR Convex Optimization 13/14

References S. Boyd and L. Vandenberghe. Convex Optimization. Berichte über verteilte messysteme. Cambridge University Press, 2004. Chuong B. Do. Convex optimization overview (cntd), 2009. J. A. Nocedal and S. J. Wright. Numerical Optimization. Springer Series in Operations Research Series. Springer-Verlag GmbH, 1999. LNMIIT MLPR Convex Optimization 14/14