A new mini-max, constrained optimization method for solving worst case problems

Similar documents
Programming, numerics and optimization

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

Computational Methods. Constrained Optimization

A Truncated Newton Method in an Augmented Lagrangian Framework for Nonlinear Programming

Computational Optimization. Constrained Optimization Algorithms

Lecture 2 Optimization with equality constraints

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Recent Developments in Model-based Derivative-free Optimization

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Introduction to Optimization Problems and Methods

Newton and Quasi-Newton Methods

3 INTEGER LINEAR PROGRAMMING

Modern Methods of Data Analysis - WS 07/08

Nonlinear Programming

The WENO Method in the Context of Earlier Methods To approximate, in a physically correct way, [3] the solution to a conservation law of the form u t

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

A Short SVM (Support Vector Machine) Tutorial

Multivariate Numerical Optimization

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

Chapter 6. Curves and Surfaces. 6.1 Graphs as Surfaces

practice: quadratic functions [102 marks]

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

A projected Hessian matrix for full waveform inversion Yong Ma and Dave Hale, Center for Wave Phenomena, Colorado School of Mines

Preface. and Its Applications 81, ISBN , doi: / , Springer Science+Business Media New York, 2013.

Composite Self-concordant Minimization

An augmented Lagrangian method for equality constrained optimization with fast infeasibility detection

Introduction to optimization methods and line search

25. NLP algorithms. ˆ Overview. ˆ Local methods. ˆ Constrained optimization. ˆ Global methods. ˆ Black-box methods.

Introduction to Optimization

Constrained and Unconstrained Optimization

The Dynamic Hungarian Algorithm for the Assignment Problem with Changing Costs

Today. Golden section, discussion of error Newton s method. Newton s method, steepest descent, conjugate gradient

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Introduction to ANSYS DesignXplorer

B553 Lecture 12: Global Optimization

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

WE consider the gate-sizing problem, that is, the problem

Topological Machining Fixture Layout Synthesis Using Genetic Algorithms

Distance-to-Solution Estimates for Optimization Problems with Constraints in Standard Form

A Brief Look at Optimization

Inverse KKT Motion Optimization: A Newton Method to Efficiently Extract Task Spaces and Cost Parameters from Demonstrations

But a vertex has two coordinates, an x and a y coordinate. So how would you find the corresponding y-value?

Convexization in Markov Chain Monte Carlo

Optimal Control Techniques for Dynamic Walking

Chapter II. Linear Programming

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

Theoretical Concepts of Machine Learning

Machine Learning for Signal Processing Lecture 4: Optimization

Convex Optimization. Lijun Zhang Modification of

OPTIMIZATION FOR AUTOMATIC HISTORY MATCHING

Using design specifications for design

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Algorithms for convex optimization

Properties of Quadratic functions

General context-free recognition in less than cubic time

OPTIMIZATION METHODS

16.410/413 Principles of Autonomy and Decision Making

Classical Gradient Methods

Short Reminder of Nonlinear Programming

Constrained Optimization COS 323

In other words, we want to find the domain points that yield the maximum or minimum values (extrema) of the function.

Solution Methods Numerical Algorithms

A robust optimization based approach to the general solution of mp-milp problems

Optimization. Industrial AI Lab.

Programs. Introduction

Scalable Algorithms in Optimization: Computational Experiments

PROJECT REPORT. Parallel Optimization in Matlab. Joakim Agnarsson, Mikael Sunde, Inna Ermilova Project in Computational Science: Report January 2013

Optimization in Scilab

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

A large number of user subroutines and utility routines is available in Abaqus, that are all programmed in Fortran. Subroutines are different for

Augmented Lagrangian Methods

for Approximating the Analytic Center of a Polytope Abstract. The analytic center of a polytope P + = fx 0 : Ax = b e T x =1g log x j

4.1 Graphical solution of a linear program and standard form

The Lucas & Kanade Algorithm

AMS527: Numerical Analysis II

Lagrangian methods for the regularization of discrete ill-posed problems. G. Landi

An Improved Measurement Placement Algorithm for Network Observability

5 Day 5: Maxima and minima for n variables.

Mathematical Programming and Research Methods (Part II)

Gate Sizing by Lagrangian Relaxation Revisited

Introduction to Linear Programming. Algorithmic and Geometric Foundations of Optimization

Distance Functions 1

Visual Tracking (1) Feature Point Tracking and Block Matching

Augmented Lagrangian Methods

Solving IK problems for open chains using optimization methods

Optimization for Machine Learning

Boundary Elements 247

IMM ROBUST NON-GRADIENT C SUBROUTINES FOR NON-LINEAR OPTIMIZATION. Pernille Brock Kaj Madsen Hans Bruun Nielsen. Revised version of report NI-91-04

SYSTEMS OF NONLINEAR EQUATIONS

MEI STRUCTURED MATHEMATICS METHODS FOR ADVANCED MATHEMATICS, C3. Practice Paper C3-B

Lecture 10: SVM Lecture Overview Support Vector Machines The binary classification problem

Motion Estimation. There are three main types (or applications) of motion estimation:

Lecture 4 Duality and Decomposition Techniques

Fast-Lipschitz Optimization

10.7 Variable Metric Methods in Multidimensions

Transcription:

Carnegie Mellon University Research Showcase @ CMU Department of Electrical and Computer Engineering Carnegie Institute of Technology 1979 A new mini-max, constrained optimization method for solving worst case problems Stephen W. Director Carnegie Mellon University Carnegie Mellon UniversityDesign Research Center Follow this and additional works at: http://repository.cmu.edu/ece This Technical Report is brought to you for free and open access by the Carnegie Institute of Technology at Research Showcase @ CMU. It has been accepted for inclusion in Department of Electrical and Computer Engineering by an authorized administrator of Research Showcase @ CMU. For more information, please contact research-showcase@andrew.cmu.edu.

NOTICE WARNING CONCERNING COPYRIGHT RESTRICTIONS: The copyright law of the United States (title 17, U.S. Code) governs the making of photocopies or other reproductions of copyrighted material. Any copying of this document without permission of its author may be prohibited by law.

A NEW MINI-MAX, CONSTRAINED OPTIMIZATION METHOD FOR SOLVING WORST CASE PROBLEMS by S.W. Director, L.M. Vidigal. R.K. Brayton** & G.P. Hachtel** DRC-18-2-79 January 1979 * Department of Electrical Engineering Carnegie-Melion University Pittsburgh, PA 15213 ** T-J. Watson Research Lab IBM Yorktown Heights, NY This research was sponsored in part by NSF Grant ENG-77-20895.

ABSTRACT A new worst case design procedure is described. This method employs Powell's new constrained optimization procedure and is at least superlinearly convergent. A novel function splitting scheme is described to avoid singularity problems inherent in some previously reported methods. I. Introduction In this paper we describe a new mini-max based method for solving both the fixed and variable tolerance worst case problems. While this method is similar to the technique proposed recently by Madsen and Schjaer- Jacobsen [1], it is at least superlinearly if not quadratically convergent even for the non regular 2 case. Furthermore, the present method solves the variable tolerance problem directly rather than as a double iteration as given in [1]. 1 2 This research was sponsored in part by NSF Grant ENG 77-20895. We are using the term regular and singular as defined for a minr-max problem in [21, i.e., a minimax problem is singular if the rank ot the Jacobian of the active constraint at the solution is less than n, the dimension of the space. University Libraries Carnegie Meifon University Pittsburgh, Pennsylvania 15213

- 2*.- The fixed tolerance problem can be stated as follows. Given the tolerance t find the nominal point x* such that the largest value of the constraints f.(x), i=l, 2,...m is minimized. Formally we can state this problem as «min max C (y) 3 x y. Rt j = i, 2,...m (1) where R = {y x.-t. _< y. _< x.4-t.} is the tolerance region. By introducing an auxiliary variable, this mini-max problem can be transformed into the constrained optimization problem. min y subject to Y _> max f j (y) Y e R t j = 1, 2,..., m. (2) In order to solve (2) (which we do using Powell's new procedure discussed in the next section) we must solve, for each function f.(y), j=l,2,...,m, the worst case problem. max f (y) y e R t Solution of the worst case problem is not easy and is discussed in Section II.

- 2JL- In the variable tolerance case the objective is to find the largest tolerances as well as the nominal design point which results in 100% yield for a feasibility region defined by the constraints f. (x) <_ c j = 1, 2,.., m. This problem can be stated as the constrained optimization problem max y subject to max f (y) < c V j = 1, 2,..., m (3) where V = {y ' X i " Yt i - y i - X i + Yt i } which is similar to (2) (note max y is the same as min -y) and can be solved using the same procedure described in Sections II and III. An example of the new algorithm is given in Section IV. II. Powell's Method Consider the problem subject to min <J>(x) (4) c ± (x) = 0, i = 1, 2,...,m where <J> and c^ are twice differentiable, real valued functions. The Kuhn- Tucker conditions for the solution of (4) are

-> 3 - m Vf(x*)- A ± * V c ± (x*) = 0 c ± (x*) =0 i - 1, 2,... m (5) If the set of equations (5) were to be solved using Newton's method we would iteratively solve the equations. m -Vc i (x k ) Ax T 7c m T (x k ) k+1 k. A k x = x +Ax = A +AA (6) Powell approximates the second derivative of the Lagrangian m L(x,A) = f(x)-x A c (x ) i«l with a positive definite matrix B that is updated at each step using a BFGS formula, and the step can be smaller than Ax in order to guarantee global convergence [5]. Also instead of solving (6) equivalent quadratic problem is solved.

- 4 - T Vf T (x )Ax k + 1/2 Ax k BAx k K. t. = 0 T k VC ro (x > (7) This problem yields the same solution as (6) and the same A! s, if B is the exact Hessian of the Lagrangian, and provides an obvious way to handle non-equality constraints, that are just append to the set of constraints after linearization. Observe that if near the solution the rank of the Jacobian of the constraints in (6) is n, the last m rows of the matrix can be used to obtain the solution, avoiding the computation of the Hessian. Because we are using Newtons method we still have 2 order convergence. This fact is why the Madsen and Schjaer-Jacobsen algorithm [1] can, in the regular case, provide a second order rate of convergence without computation of second derivatives [3]. But more important is the observation that we should try to introduce the largest number of possible constraints in order to improve the rate of convergence of the method. This observation is the motivation behind the concept of function splitting which is introduced in the next section. Note that as with any Quasi-Newton optimization method at least superlinear convergence should be expected [6J.

- 5 - III. The Worst Case Problem As was pointed in Section 1, in order to solve (2) we have to solve the worst case subproblera max fj y c R t (8) for each j = 1, 2,..., m. In general this problem is difficult to solve, and some restrictions need to be made on the f.. In the following development we will assume that the f. are such that the solution of (8) is always found at one vertex of R. Even with this simplification, we cannot use Powell's method directly because the constraints in (2) are not smooth. A small variation of the centers of the tolerance region can cause a change of the worst case vertex, or the maximum value can be achieved at more than one vertex simultaneousely. To solve this problem at each iteration we not only compute the worst case vertex but also introduce new inequalities corresponding to the vertices that have been active in one of the n previous iterations. As was mentioned in Section II, introduction of new constraints can only help the rate of convergence of the algorithm. In the example of Figure 1, where m = 1, at the first iteration we would have only one constraint corresponding to vertex V2, but in all subsequent iterations we would always have 2 inequalities, for vertex V~ and V~. Geometrically it can be seen that it prevents the algorithm from bouncing around the 45 degree line. In general, any function could be split into n+1 constraints,but as will be seen in the examples this situation is very unlikely. Another way of looking to the algorithm is as follows. Let R (x) represent the set of vertices of the tolerance region corresponding to a

6 - nominal value x. Under the assumptions made above, (2) can be recast as nin y j = 1,.., m i = 1,..., 2 n (7) y i e R V (x) Notice that the number of constraints would be mx2 making the computation prohibitive for large values of n. The proposed method is basically a scheme to pick out those constraints which are good candidates to be active. IV. Examples constraint Consider as a first example the fixed tolerance problem (1) with one f(x) =.505x 1 2 +.505x 2 2 - -99x^2 with the tolerances t = (0.1, 0.1). The level cuts are elongated ellipses and the minimum of f(x) is obtained at JC = (0.,0.) Note that since we have one constraint and two variables (i.e., m=l and n=2) this problem is singular even with constraint splitting. Starting with the point x = (2,A) and Y = l> the algorithm took 13 iterations to converge to the solution x = (0,0) requiring 66 function evaluations and 55 gradients, The variable tolerance problem, with c= 0.1 with the same starting point that took 12 iterations to converge to the solution x = (0,0), r = 2.236 requiring 63 functions evaluations and 51 gradients.

- 7 - When solving the subproblem max f,(y) we had a problem whenever yer t x =x 9. The algorithm we use to find y is based upon gradient information. The gradient at the center of R is computed and we take the vertex pointed by the signs of this gradient. If the signs of the gradient are the same at this vertex a maximum was found, otherwise we take the vertex pointed by the new gradient and the procedure is repeated. If x^x^ it can easily be checked that the maximum found is a local maximum. It was decided that before taking a vertex given by this procedure as the worst case point, we always check the function values at the worst vertices in the last n iterations. Notice that no extra function evaluations are required because the function at these vertices need to be computed anyway due to the splitting mechanism. For comparison we tried the above algorithm to solve example 1 of [1]- Both fixed tolerance problem and variable tolerance problem took only 5 iterations to converge to the solution. For functions f ~ and f~ only one vertex becomes active. For function f-.,2 vertices are active at the solution.

References [1] K. Madsen and H. Schjaer Jacobsen, "Algorithms for Worst Case Tolerance Optimization." Report NI-78-05, Institute of Numerical Analysis, Technical University of Denmark. [2], "Singularities in Minimax Optimization of Networks," IEEE Transactions on CAS, Vol. CAS-23, No. 7, July 1974, pp 456-560. [3]., "Linearly Constrained Minimax Optimization," Mathematical Programming, 14 (1978) 208-223, [4] M. J. D. Powell, "Algorithms for Nonlinear Constraints That Use Langrangian Functions," Mathematical Programming 14 (1978) 224-248. f 5 -! M. J. D. Powell, "A Past Algorithm for Nonlinear Constrained Optimization Calculations", Presented at the 1977 Dundee Conference on Numerical Analysis. [6] M. J. D. Powell, "The Convergence of Variable Metric Methods for Nonlinear Constrained Optimization Calculations."

FIG. l EXAMPLE 1 X1

f(x) x7 / /" -3 FIG. 2 EXAMPLE 1 ( c=.1 )

2.4.6.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 FIG. 3 EXAMPLE 2 (f.1 t.1)

2.2 2.0 1.8 1.6 1.4 1.2 1.0.8.6.4.0.2 A.6.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 FIG. 4 EXAMPLE 2 ( c-1.5 )