Computational Optimization. Constrained Optimization Algorithms

Similar documents
Computational Optimization. Constrained Optimization

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Computational Methods. Constrained Optimization

Programming, numerics and optimization

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Local and Global Minimum

Introduction to Optimization Problems and Methods

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Chapter 3 Numerical Methods

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

Lecture Notes: Constraint Optimization

Introduction to Optimization

Characterizing Improving Directions Unconstrained Optimization

Constrained Optimization COS 323

Lecture 12: Feasible direction methods

Demo 1: KKT conditions with inequality constraints

Machine Learning for Signal Processing Lecture 4: Optimization

Convex Optimization and Machine Learning

A FACTOR GRAPH APPROACH TO CONSTRAINED OPTIMIZATION. A Thesis Presented to The Academic Faculty. Ivan Dario Jimenez

Optimization III: Constrained Optimization

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

CS231A. Review for Problem Set 1. Saumitro Dasgupta

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

A Short SVM (Support Vector Machine) Tutorial

Introduction to optimization methods and line search

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year

Augmented Lagrangian Methods

Lecture 18: March 23

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Lecture 12: convergence. Derivative (one variable)

Optimal Control Techniques for Dynamic Walking

Augmented Lagrangian Methods

Convex Optimization CMU-10725

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.

5 Machine Learning Abstractions and Numerical Optimization

Lecture 2 Optimization with equality constraints

(1) Given the following system of linear equations, which depends on a parameter a R, 3x y + 5z = 2 4x + y + (a 2 14)z = a + 2

Programs. Introduction

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

Mathematical Programming and Research Methods (Part II)

ISM206 Lecture, April 26, 2005 Optimization of Nonlinear Objectives, with Non-Linear Constraints

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

Department of Electrical and Computer Engineering

Optimization under uncertainty: modeling and solution methods

Least-Squares Minimization Under Constraints EPFL Technical Report #

A new mini-max, constrained optimization method for solving worst case problems

Solution Methods Numerical Algorithms

Parallel and Distributed Graph Cuts by Dual Decomposition

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

1) Give a set-theoretic description of the given points as a subset W of R 3. a) The points on the plane x + y 2z = 0.

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions

Fast-Lipschitz Optimization

Short Reminder of Nonlinear Programming

Optimization. Industrial AI Lab.

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?

Interior Point I. Lab 21. Introduction

4.1 Graphical solution of a linear program and standard form

Lecture 9. Introduction to Numerical Techniques

Constrained optimization

CS 6210 Fall 2016 Bei Wang. Review Lecture What have we learnt in Scientific Computing?

Lecture 19: Convex Non-Smooth Optimization. April 2, 2007

Distributed Estimation for Spatial Rigid Motion Based on Dual Quaternions

5 Day 5: Maxima and minima for n variables.

Introduction to Optimization

Modern Methods of Data Analysis - WS 07/08

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi

21-256: Lagrange multipliers

16.410/413 Principles of Autonomy and Decision Making

OPTIMIZATION FOR AUTOMATIC HISTORY MATCHING

Computational Methods CMSC/AMSC/MAPL 460. Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science

Today. Golden section, discussion of error Newton s method. Newton s method, steepest descent, conjugate gradient

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

Finite Math - J-term Homework. Section Inverse of a Square Matrix

Hartley - Zisserman reading club. Part I: Hartley and Zisserman Appendix 6: Part II: Zhengyou Zhang: Presented by Daniel Fontijne

Convex Optimization MLSS 2015

Search direction improvement for gradient-based optimization problems

A Numerical Study of Active-Set and Interior-Point Methods for Bound Constrained Optimization

Linear methods for supervised learning

Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem

AGloballyConvergentPrimal-DualActive-Set Framework for Large-Scale Convex Quadratic Optimization

Theoretical Concepts of Machine Learning

x = 12 x = 12 1x = 16

A primal-dual framework for mixtures of regularizers

Convexity Theory and Gradient Methods

Composite Self-concordant Minimization

Revision Problems for Examination 2 in Algebra 1

Classical Gradient Methods

Optimization Algorithms, Implementations and. Discussions (technical report for self-reference)

Applied Lagrange Duality for Constrained Optimization

3D Geometry and Camera Calibration

In other words, we want to find the domain points that yield the maximum or minimum values (extrema) of the function.

Full waveform inversion by deconvolution gradient method

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding

Chapter II. Linear Programming

Transcription:

Computational Optimization Constrained Optimization Algorithms

Same basic algorithms Repeat Determine descent direction Determine step size Take a step Until Optimal But now must consider feasibility, e.g. Pick a feasible descent directions Choose step size to maintain feasibility Need dual multipliers to check optimality

Builds on Prior Approaches Unconstrained: min f ( x) Lots of approaches Linear Equality Constrained: min f ( x) st.. Ax= b Convert to unconstrained and solve

Prior Approaches (cont) Linear Inequality Constrained: min f ( x) st.. Ax b Identify active constraints Solve subproblems Nonlinear Inequality Constrained: min f( x) st.. g( x) 0 Linearize constraints Solve subproblems

Feasible Direction Methods Use feasible descent directs. Consider 1 min xqx ' + cx ' 2 st.. Ax= b If we have feasible point ˆx then all x = xˆ + Zp are feasible.

Reduced Problem 1 min p ( x ˆ + Zp) ' Q( x ˆ + Zp) + c'( x ˆ + Zp) 2 1 1 min ' ' ( ˆ ) ' ˆ ˆ ' ˆ p p Z QZp + Qx + c Zp + xqx + c x 2 2 1 min ' ' ( ˆ p pzqzp+ Qx+ c) ' Zp 2 ZQZ ' : = Reduced Hessian Z'( Q( xˆ + Zp) + c) : = Reduced Gradient

Reduced Newton s Equation For general problems Z 2 ' f( x) Z : = Reduced Hessian Z' f( x) : = Reduced Gradient Reduced Newton Equations Z f x Z v Z f x 2 ' ( ) = - ' ( ) v = Z f x Z Z f x 2 1 ( ' ( ) ) ' ( ) in reduced space 2 1 p = Zv = Z( Z' f( x) Z ) Z' f( x) in original space Then apply usual Newton s Method

Reduced Steepest Descent Use reduced steepest descent direction v = Z' f( x) in reduced space d = Zv = ZZ' f( x) in original space Convergence rate depends on condition number of reduced Hessian Cond Z f x Z Cond Z Z Cond f x 2 2 ( ' ( ) ) ( ' ) ( ( )) KEY POINT Construct Z to have good conditioning!

Ways to Compute Z Projection Method (orthogonal and non-orthogonal) Variable Reduction QR Factorization What are they and how do they effect conditioning? How can you use them to compute multipliers?

Projection Method Find closest feasible point 1 2 2 min c x Ax = 0 x Compute a KKT point: This is optimal projection since problem is convex

min c x Ax = 0 x 1 2 1 2 Projection Method Lx (, λ) = c x + ( Ax)' λ ( ) xlx (, λ) = c x + A' λ = 0 KKT are Ax b = 0 Ac Ax AA' λ = 0 ( ') 1 ( ') ( ') 1 AA' λ = Ac λ = AA Ac x = c A' AA Ac 1 x = I A' AA A c ( ) 1 Projection Matrix P = I A' AA' A

Project method The matrix ( ) ' 1 Z = I A' AA A is a basis for the null space matrix of A. Check ( ) AZ = A I A' AA' 1 A = A A = 0

Get Lagrangian Multipliers for free! The matrix 1 1 A A' AA' = where AA = AA' AA' = I r ( ) ( ) is the right inverse matrix for A. For general problems r min f ( x) st.. Ax λ ' r * = A f( x*) = b

Let s try it For 1 2 2 2 2 min f( x) = 2 x + x + x + x 1 2 3 4 st.. x + x + x + x = 1 Projection matrix 1 2 3 4 1 0 0 0 1 1 0 1 0 0 1 1 Z = I A' AA' A= 1 1 1 1 1 1 1 1 0 0 1 0 1 1 0 0 0 1 1 1 1 ( ) [ ] [ ] 3 1 1 1 4 4 4 4 1 3 1 1 4 4 4 4 = 1 1 3 1 4 4 4 4 1 1 1 3 4 4 4 4 1

Solve FONC for Optimal Point FONC x 1 1 x 1 f( x) A' λ λ x 1 x 1 4 x + x + x + x = 1 2 = = 1 2 3 4

Check Optimality Conditions For x* = [1111]/ 4 f( x*) = [1111]/ 4 Ax* = b 3 1 1 1 4 4 4 4 1 1 3 1 1 1 4 4 4 4 1 Z f(*) x = * = 0 1 1 3 1 1 4 4 4 4 4 1 1 1 3 1 4 4 4 4 Using Lagrangian ( ) 1 1 Ar = A' AA' = [1111]' 4 λ = A f(*) x = 1/4 r Clearly f(*) x = A' λ

For min f( x) = x' Cx 1 2 st.. Ax= b You try it 0 13 6 3 13 23 9 3 2 1 2 1 2 C = A= b = 6 9 12 1 1 1 3 1 3 3 3 1 3 Find projection matrix Confirm optimality conds are Z Cx*=0, Ax* = b Find x* Compute Lagrangian multipliers Check Lagrangian form of the multipliers.

Projected Gradient Method Algorithm for min f(x) s.t. Ax=b Let Z=I-A (AA ) -1 A x 0 given g =gradf(x 0 ) While not optimal d=-z g x i+1 = x i+ αd (use linesearch) g =gradf(x i+1 ) i=i+1

Projected Gradient Method Equivalent to steepest descent on f(x+zp) where x is some feasible point. So convergence rate depends on condition number of this problem The condition number is less than equal condition of Hessian of f * Compute Z HZ for last example and check it s conditioning:

Lagrangian Multipliers Each null space method yields a corresponding calculation of the right inverse. Thus each yields calculation of corresponding dual multipliers.

Variable Reduction Method Let A=[B N] A is m by n B is m by m assume m < n Z 1 B N = I is a basis matrix for null space of A 1 1 B B A [ ] r = AAr = BN = I + 0 0 0

Try on our example Take for example first two columns for B Then 2 1 2 1 2 1 2 1 A= = [ B N] = 1 1 3 1 1 1 3 1 1 2 1 1 4 3 1 2 Z = A r = 1 0 0 0 0 1 0 0 Condition number of Z CZ = 158 better but not great

QR Factorization Use Gram-Schmidt algorithm to make orthogonal factorize A =QR with Q orthogonal and R upper triangular R1 A' = QR = [ Q1 Q2] 0 where A m n, Q n m, Q n ( n m), R m m Z = Q A = QR T 2 r 1 1 1 2 1

QR on problem Use matlab command QR [Q R ] = qr(a ) Q2 = Q(:,3:4) Cond(Q2 *C*Q2) = 9.79

Next Problem Consider Next Hardest Problem min f ( x) s.. t Ax b How could we adapt gradient projection or other linear equality constrained problem to this problem?