Computational Methods. Constrained Optimization

Similar documents
Programming, numerics and optimization

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

Characterizing Improving Directions Unconstrained Optimization

Chapter II. Linear Programming

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

16.410/413 Principles of Autonomy and Decision Making

Introduction to Optimization Problems and Methods

Introduction to Optimization

Chapter 3 Numerical Methods

Constrained and Unconstrained Optimization

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Introduction to Optimization

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

Constrained Optimization COS 323

A FACTOR GRAPH APPROACH TO CONSTRAINED OPTIMIZATION. A Thesis Presented to The Academic Faculty. Ivan Dario Jimenez

Today. Golden section, discussion of error Newton s method. Newton s method, steepest descent, conjugate gradient

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

Computational Optimization. Constrained Optimization Algorithms

The Simplex Algorithm

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year

Mathematical Programming and Research Methods (Part II)

Machine Learning for Signal Processing Lecture 4: Optimization

Nonlinear Programming

Local and Global Minimum

Lecture 12: Feasible direction methods

Modern Methods of Data Analysis - WS 07/08

5 Machine Learning Abstractions and Numerical Optimization

Algorithms for convex optimization

CMU-Q Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization. Teacher: Gianni A. Di Caro

Solution Methods Numerical Algorithms

5 Day 5: Maxima and minima for n variables.

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?

Convex Optimization CMU-10725

Lecture 2 Optimization with equality constraints

A Short SVM (Support Vector Machine) Tutorial

Optimization III: Constrained Optimization

Convexization in Markov Chain Monte Carlo

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

(1) Given the following system of linear equations, which depends on a parameter a R, 3x y + 5z = 2 4x + y + (a 2 14)z = a + 2

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Handling of constraints

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):

Linear Programming. them such that they

Demo 1: KKT conditions with inequality constraints

Introduction to optimization methods and line search

Fast-Lipschitz Optimization

Lecture 2 - Introduction to Polytopes

21-256: Lagrange multipliers

ISM206 Lecture, April 26, 2005 Optimization of Nonlinear Objectives, with Non-Linear Constraints

Lecture Notes 2: The Simplex Algorithm

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi

3 INTEGER LINEAR PROGRAMMING

Convexity Theory and Gradient Methods

we wish to minimize this function; to make life easier, we may minimize

Data Analysis 3. Support Vector Machines. Jan Platoš October 30, 2017

Linear Programming Problems

Lecture 15: Log Barrier Method

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Recent Developments in Model-based Derivative-free Optimization

SYSTEMS OF NONLINEAR EQUATIONS

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Lecture 18: March 23

Constrained Optimization with Calculus. Background Three Big Problems Setup and Vocabulary

Distributed Alternating Direction Method of Multipliers

Optimization under uncertainty: modeling and solution methods

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

B553 Lecture 12: Global Optimization

Lagrange multipliers October 2013

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Mathematical and Algorithmic Foundations Linear Programming and Matchings

25. NLP algorithms. ˆ Overview. ˆ Local methods. ˆ Constrained optimization. ˆ Global methods. ˆ Black-box methods.

The Alternating Direction Method of Multipliers

4.1 Graphical solution of a linear program and standard form

Kernels and Constrained Optimization

minimise f(x) subject to g(x) = b, x X. inf f(x) = inf L(x,

MATH3016: OPTIMIZATION

Kernel Methods & Support Vector Machines

Theoretical Concepts of Machine Learning

Programs. Introduction

CSE 460. Today we will look at" Classes of Optimization Problems" Linear Programming" The Simplex Algorithm"

Convex Optimization and Machine Learning

5.3 Cutting plane methods and Gomory fractional cuts

CSC 8301 Design & Analysis of Algorithms: Linear Programming

EC5555 Economics Masters Refresher Course in Mathematics September Lecture 6 Optimization with equality constraints Francesco Feri

Integer Programming Theory

MATH Lagrange multipliers in 3 variables Fall 2016

UNIT 3 EXPRESSIONS AND EQUATIONS Lesson 3: Creating Quadratic Equations in Two or More Variables

Lagrange multipliers 14.8

for Approximating the Analytic Center of a Polytope Abstract. The analytic center of a polytope P + = fx 0 : Ax = b e T x =1g log x j

CasADi tutorial Introduction

LECTURE 6: INTERIOR POINT METHOD. 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm

Multivariate Numerical Optimization

x 6 + λ 2 x 6 = for the curve y = 1 2 x3 gives f(1, 1 2 ) = λ actually has another solution besides λ = 1 2 = However, the equation λ

1 Linear programming relaxation

Introduction to optimization

EARLY INTERIOR-POINT METHODS

Transcription:

Computational Methods Constrained Optimization Manfred Huber 2010 1

Constrained Optimization Unconstrained Optimization finds a minimum of a function under the assumption that the parameters can take on any possible value. In a range of problems additional constraints exist that limit the range of feasible parameters. Unconstrained optimization techniques would often find solutions that are not feasible (parameters do not fulfill constraints) Simply limiting iterative methods at the constraints will often lead to suboptimal solutions Manfred Huber 2010 2

Constrained Optimization Constrained optimization problems can be defined using an objective function and a set of constraints. Objective function: min x f(x) Equality constraints: g i (x)=0 Inequality constraints: h i (x) 0 A feasible point is any point that fulfills all the constraints. An optimal point is one that locally optimizes the value function given the constraints. Manfred Huber 2010 3

Constrained Optimization A number of special cases for constrained optimization problems can be defined based on the types of functions used Linear programming: Linear objective function: f(x)= T x Linear constraints: g(x)=bx-c ; h(x)=cx-d Quadratic programming: Quadratic objective function: f(x)=1/2 x T Ax + T x Linear constraints: g(x)=bx-c ; h(x)=cx-d Manfred Huber 2010 4

Constrained Optimization Solving constrained optimization problems is substantially harder than solving unconstrained optimization problems If possible, convert the constrained optimization problem into an unconstrained optimization problem For constraints on individual parameters, replace them with differentiable functions of an unconstrained parameter (b a) a x i b x i = 1+ e x i x i a x i = e x i + a x i + αx j = 0 x i = αx j + a Manfred Huber 2010 5

Constrained Optimization Another differentiation can be made in terms of the types of constraints present Optimization problems with only equality constraints Optimization problems with inequality constraints Equality constraints are generally easier to deal with than inequality constraints Manfred Huber 2010 6

Problems with Equality Constraints A problem with only equality constraints is defined through and objective function f(x) and a set of constraints, g(x)=0 Feasible points have to fulfill the constraints Constrained optimal points have to fulfill the necessary condition that the negative gradient of the objective function can not have any component that falls within the space spanned by the constraints Otherwise there would be a way to lower the objective function without violating the constraints f (x * ) = J g T (x * )λ Manfred Huber 2010 7

Problems with Equality Constraints Lagrange multipliers introduce the equality constraints into the optimality condition Instead of optimizing f(x) optimize the Lagrangian function L(x,) L(x,λ) = f (x) + λ T g(x) At the constrained minimum the Lagrangian has a minimum and therefore constrained optimization solution has to be a solution to the nonlinear system L(x,λ) = f (x) + J T g (x)λ = 0 g(x) Manfred Huber 2010 8

Equality Constraint Optimization Problems with equality constraints can be solved by finding a critical point of the Lagrangian using Newton s method L(x,λ) = f (x) + J T g (x)λ = 0 g(x) To find this point requires the Hessian H L (x,λ) = B(x,λ) J T g (x) J g (x) 0 B(x,λ) = H f (x) + m i=1 H gi (x) Manfred Huber 2010 9

Equality Constraint Optimization The step for Newton s method can be determined (as generally in Newton s method) as the solution to the linear system B(x,λ) J T g (x) s J g (x) 0 δ = f (x) + J T (x)λ g g(x) In the case of pure equality constraints this can be solved directly by solving the (m+n)x(m+n) linear system Solution has to be verified as a minimum since it could be a saddle Manfred Huber 2010 10

Equality Constraint Optimization We need a means to judge the progress to evaluate convergence and feasibility x in each step does not necessarily fulfill the constraints Only when converged do the constraints have to be fulfilled Merit functions can be used to measure progress E.g. Penalty function φ p (x) = f (x) + 1 2 ρg(x)t g(x) represents tradeoff between optimality and feasilbity Penalty function for a large has the same minimum as the original function Manfred Huber 2010 11

Equality Constraint Optimization Penalty function provides another means to solve an equality constrained optimization problem Constrained minimum of f(x) is the same as the unconstrained minimum of the penalty function for a sufficiently large Penalty function allows to convert a constrained optimization problem into an unconstrained problem Unconstrained problem becomes ill-conditioned for large Penalty term starts to dominate and f(x) disappears Range from which optimization converges correctly becomes small Use of iterative approach can solve problem Incrementally solve unconstrained optimization problems for increasing values of Manfred Huber 2010 12

Inequality Constraints With inequality constraints, the matrix for Newton s method can no longer be solved directly. Sometimes inequality constraints can be (partially) converted into equality constraints Active set methods replace inequality constraints with equality constraints for a subset of the constraints Inequality constraints that are not currently violated can be temporarily dropped as the next step is unlikely to violate them Inequality constraints that are violated can be replaced by equality constraints since the equality is the closes value to fulfilling the constraint Active set method requires to reevaluate the set of required constraints at each step Manfred Huber 2010 13

Extended Form Inequality constraints can be replaced by equality constraints and an additional, simpler inequality constraint Introducing slack variables h(x) 0 h(x) + x s = 0, x s 0 System with slack variables can be represented as an extended system (m h additional variables) with solely constraints on the slack variables Allows more efficient solution methods since all constraints are uniform Manfred Huber 2010 14

Barrier Methods As for equality constraints, optimization problems with inequality constraints can be solved by converting them into a sequence of unconstrained optimization problems with appropriate penalties Barrier functions form one-sided penalty function m 1 φ µ (x) = f (x) µ p i=1 h i (x) For sufficiently small the barrier function has the same minimum as the constrained optimization problem for f(x) Again ill conditioned for small Iterated unconstrainted optimization using barrier function with decreasing can address problem Basis for interior point methods Manfred Huber 2010 15

Sequential Quadratic Programming Lagrange multipliers can also be used with inequality constraints but have to be 0 for all inactive constraints Instead of optimizing f(x) we can again optimize the Lagrangian function L(x,) At the constrained minimum the Lagrangian has a minimum and therefore constrained optimization solution has to be a solution to the nonlinear system L(x,λ) = f (x) + J T g (x)λ = 0 g(x) Manfred Huber 2010 16

Sequential Quadratic Programming The step for Newton s method can be determined (as generally in Newton s method) as the solution to the linear system with the added constraints for active set variables B(x,λ) J T g (x) s J g (x) 0 δ = f (x) + J T (x)λ g g(x) This can no longer be solved directly but represents a quadratic programming problem for f q (s) = 1 2 st B(x,λ)s + s T ( f (x) + J T g,h (x)λ) J g (x)s + g(x) = 0 J h (x)s + h(x) 0 Manfred Huber 2010 17

Linear Programming Linear programs are an important special class of constrained optimization problems Linear objective function f(x) Linear constraints g(x), h(x) Often with the additional constraints x 0 In linear programming problems (as in quadratic programming problems), the feasible region is convex Since objective function is linear the optimum has to occur at a vertex of the constrained feasible region Solution is at an intersection point of n constraints Simplex algorithm uses extended form to navigate a search through the space of vertices Manfred Huber 2010 18

Simplex Algorithm In augmented form a vertex is characterized as a point in which at least n variables are 0 Z 1 c T 0 x 0 A I = 0, x 0 x b s 0 Starting from one vertex, navigate to a neighbor with a lower value of Z x s Neighbors are nodes in which one of the n variables that were set to 0 become non-zero Selection of neighbor is the one that requires the smallest change (i.e. along steepest gradient ) Manfred Huber 2010 19

Simplex Algorithm In the worst case the complexity of the simplex algorithm is exponential Problems can be constructed where it moves through every single vertex In practice it is usually very efficient and most of the time outperforms interior point methods (which are polynomial) Simplex algorithm can solve problems with large numbers of variables Manfred Huber 2010 20

Manfred Huber 2010 21 Simplex Algorithm - Example Linear program: Objective function: P=-2x-3y-4z Constraints: 3x+2y+z 10, 2x+5y+3z 15, x,y,z 0 Convert linear program into extended form: Objective function: P+2x+3y+4z=0 Constraints: 3x+2y+z+s=10, 2x+5y+3z+t=15, x,y,z,s,t 0 1 2 3 4 0 0 0 3 2 1 1 0 0 2 5 3 0 1 P x y z s t = 0 10 15

Simplex Algorithm - Example Determine initial vertex by identifying 3 variables that can simultaneously be 0 under the constraints Convert the matrix such that first row contains 0 for all elements except the first column and the 3 variables that are equal to 0 In this case this is already the case when selecting x, y, z. Thus starting with the vertex described by x=y=z=0 To find next vertex, find the variable among the current ones that has the steepest downhill gradient in the objective function (i.e. highest coefficient in the first row) in this case z P 1 2 3 4 0 0 x 0 y 0 3 2 1 1 0 = z 10 0 2 5 3 0 1 15 s t Manfred Huber 2010 22

Simplex Algorithm - Example Identify the next variable to set to 0 (instead of z) by finding the constraint row that requires the minimal change in z (computed by dividing the right hand side of the constraint equations by the corresponding value in the z column and then selecting the constraint equation with the smallest value in this case the second constraint (third row) and thus the variable t P 1 2 3 4 0 0 x 0 y 0 3 2 1 1 0 = z 10 0 2 5 3 0 1 15 s t 10/1=10 15/3=5 Manfred Huber 2010 23

Simplex Algorithm - Example Eliminate the entries in the z column for all but the selected constraint row and eliminate all entries corresponding to non-zero variables in the first row. Find the direction towards the next vertex by identifying the strongest downhill direction for the selected variables (x,y,t) P 1 2 3 11 3 0 0 4 3 x 20 y 0 8 3 1 3 0 1 1 3 = z 5 0 2 5 3 0 1 15 s t Since all entries in the first row are negative, there is now way to improve the function and we have reached the optimal vertex with x=y=t=0 Manfred Huber 2010 24

Simplex Algorithm - Example Find the values for all the variables in the vertex P 1 2 3 11 3 0 0 4 3 x 20 y 0 8 3 1 3 0 1 1 3 = z 5 0 2 5 3 0 1 15 s t From the first row we obtain the objective function P=2/3x+11/3y+4/3t-20 which is (as it has to be) minimal at x=y=t=0 with a value of -20 Using the second constraint equation and the already known variable values we obtain 2x+5y+3z+t = 3z = 15 and thus z=5 Solution is x=y=0, z=5, P=-20 Manfred Huber 2010 25

Constrained Optimization Constrained Optimization allows to find the best parameters for arbitrary objective functions while obeying constraints Equality and inequality constraints Special cases of Linear programming can be addressed using the constraints explicitly Simplex algorithm Most general solution methods transform the constrained problem into an unconstrained one with the same solution Lagrangian function Penalty or barrier functions Convergence is enhanced by iterating over increasingly strict barrier functions Manfred Huber 2010 26