EARLY INTERIOR-POINT METHODS

Similar documents
Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Convex Optimization CMU-10725

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

LECTURE 6: INTERIOR POINT METHOD. 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?

CMPSCI611: The Simplex Algorithm Lecture 24

Programming, numerics and optimization

Mathematical and Algorithmic Foundations Linear Programming and Matchings

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

George B. Dantzig Mukund N. Thapa. Linear Programming. 1: Introduction. With 87 Illustrations. Springer

Linear Programming 1

February 19, Integer programming. Outline. Problem formulation. Branch-andbound

16.410/413 Principles of Autonomy and Decision Making

Lecture 16 October 23, 2014

Linear Programming Duality and Algorithms

Solution Methods Numerical Algorithms

Linear Optimization and Extensions: Theory and Algorithms

Discrete Optimization. Lecture Notes 2

THEORY OF LINEAR AND INTEGER PROGRAMMING

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

3 Interior Point Method

Convex Optimization M2

15. Cutting plane and ellipsoid methods

Algorithms for convex optimization

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Chapter 15 Introduction to Linear Programming

Introduction to Linear Programming. Algorithmic and Geometric Foundations of Optimization

Math 5593 Linear Programming Lecture Notes

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming

Nonlinear Programming

CS 473: Algorithms. Ruta Mehta. Spring University of Illinois, Urbana-Champaign. Ruta (UIUC) CS473 1 Spring / 50

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini

The Simplex Algorithm

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

Selected Topics in Column Generation

Linear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage).

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Introduction to Optimization Problems and Methods

(67686) Mathematical Foundations of AI July 30, Lecture 11

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.

Mathematical Programming and Research Methods (Part II)

Integer Programming Theory

INTRODUCTION TO LINEAR AND NONLINEAR PROGRAMMING

Applied Lagrange Duality for Constrained Optimization

VARIANTS OF THE SIMPLEX METHOD

MA4254: Discrete Optimization. Defeng Sun. Department of Mathematics National University of Singapore Office: S Telephone:

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs

Linear Programming: Mathematics, Theory and Algorithms

B553 Lecture 12: Global Optimization

Linear Programming and its Applications

Algorithmic Game Theory and Applications. Lecture 6: The Simplex Algorithm

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 1: Introduction to Optimization. Instructor: Shaddin Dughmi

16.410/413 Principles of Autonomy and Decision Making

Notes for Lecture 18

CSC 8301 Design & Analysis of Algorithms: Linear Programming

Lecture 2 - Introduction to Polytopes

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

EE/AA 578: Convex Optimization

A primal-dual Dikin affine scaling method for symmetric conic optimization

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

5 Machine Learning Abstractions and Numerical Optimization

ORIE 6300 Mathematical Programming I November 13, Lecture 23. max b T y. x 0 s 0. s.t. A T y + s = c

Computational Methods. Constrained Optimization

CS 473: Algorithms. Ruta Mehta. Spring University of Illinois, Urbana-Champaign. Ruta (UIUC) CS473 1 Spring / 36

AN EXPERIMENTAL INVESTIGATION OF A PRIMAL- DUAL EXTERIOR POINT SIMPLEX ALGORITHM

Chapter II. Linear Programming

Lecture 2 September 3

Lecture 12: Feasible direction methods

LARGE SCALE LINEAR AND INTEGER OPTIMIZATION: A UNIFIED APPROACH

Section Notes 5. Review of Linear Programming. Applied Math / Engineering Sciences 121. Week of October 15, 2017

1. Introduction. performance of numerical methods. complexity bounds. structural convex optimization. course goals and topics

MATH3016: OPTIMIZATION

1 Linear programming relaxation

A PRIMAL-DUAL EXTERIOR POINT ALGORITHM FOR LINEAR PROGRAMMING PROBLEMS

Size of a problem instance: Bigger instances take

Convex Programs. COMPSCI 371D Machine Learning. COMPSCI 371D Machine Learning Convex Programs 1 / 21

Simplex Algorithm in 1 Slide

Some Advanced Topics in Linear Programming

Numerical Robustness. The implementation of adaptive filtering algorithms on a digital computer, which inevitably operates using finite word-lengths,

Chapter 4 Concepts from Geometry

CS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm. Instructor: Shaddin Dughmi

6.854 Advanced Algorithms. Scribes: Jay Kumar Sundararajan. Duality

Lecture 14: Linear Programming II

Outline. Column Generation: Cutting Stock A very applied method. Introduction to Column Generation. Given an LP problem

Column Generation: Cutting Stock

Lecture Notes 2: The Simplex Algorithm

Linear Programming: Introduction

Algorithms for Integer Programming

J Linear Programming Algorithms

maximize c, x subject to Ax b,

Math 273a: Optimization Linear programming

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Introduction to Linear Programming

GENERAL ASSIGNMENT PROBLEM via Branch and Price JOHN AND LEI

Optimization Methods for Machine Learning (OMML)

Computing intersections in a set of line segments: the Bentley-Ottmann algorithm

A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems

3 INTEGER LINEAR PROGRAMMING

1.2 Numerical Solutions of Flow Problems

Transcription:

C H A P T E R 3 EARLY INTERIOR-POINT METHODS An interior-point algorithm is one that improves a feasible interior solution point of the linear program by steps through the interior, rather than one that improves by steps around the boundary of the feasible region, as the classical Simplex Algorithm does. The earliest interior-point method is due to the famous mathematician John von Neumann. His method for finding a feasible solution to a linear program with a convexity constraint is notable for its simplicity and remarkable convergence properties; see Section 3.1. Since a general linear program combined with its dual can be reformulated into a feasibility problem of this restricted form, von Neumann s algorithm may be viewed as a method for solving the general linear program. Just like there are many variants of the Simplex Method (which we refer to as pivot step algorithms), so there are many variants of interior methods such as projective and/or potential reduction, affine, andpath-following. 1. Projective and Potential Reduction Methods. These methods measure the approach toward an optimal solution by the reduction of the value of a potential function rather than the reduction of the value of the linear objective. For example, Karmarkar s agorithm is typically based on projective geometry but uses a potential function to measure progress of the solution towards optimality. The potential function is typically designed to ensure the following: (a) the objective function decreases at each iteration, (b) the solution point stays in the interior of the feasible space, and (c) the algorithm converges in polynomial time. In practice, these methods have not done well. 2. Affine Methods. These methods approximate the feasible region, at each iteration, by an ellipsoid and optimize over the ellipsoid. The implementation of 67

68 EARLY INTERIOR-POINT METHODS such methods is easy as we saw in the discussion of one such method in Linear Programming 1: Introduction. In this chapter, we discuss Dikin s method, an early affine method. In practice these methods perform quite well but not as well as the path-following methods. 3. Path-Following Methods. These methods follow a certain path as the optimal solution is approached. The linear program is first transformed into an unconstrained nonlinear optimization problem, called a logarithmic barrier function. The logarithmic barrier function typically consists of the objective function and one or more additional terms, multiplied by a scalar positive parameter, that increase in value as the iterates approach the boundary. In effect, the additional terms throw up a barrier at the boundary. The unconstrained optimization problem is solved and the parameter value reduced for the next iteration. The optimal values of the sequence of unconstrained problems approach the optimal solution of the linear program along a path through the interior of the feasible region. Path-following methods have performed the best in theory and practice in recent times. In Chapter 4 we will describe the primal logarithmic barrier method and the primal-dual logarithmic barrier method. Some other interior-methods inscribe an ellipsoidal ball in the feasible region with its center at the current iterate, or first transform the feasible space and then inscribe a hypersphere with the current iterate at the center. Then an improving direction is found by joining the current iterate to the point on the boundary of the ellipsoid or sphere that maximizes (or minimizes) the linear objective function (obtained by solving a least-squares problem). A point is then selected on the improving direction line as the next current iterate. Sometimes this iterate is found along a line that is a linear combination of the improving direction and some other direction. In 1967 Dikin proposed an affine method that in its original form is not a finite method but one that converges in the limit. In particular, Dikin s method as described in Section 3.2, has an asymptotic rate of convergence of 1 1/ m. This method has the distinction of having been rediscovered by many; for example, the primal affine method is the same as Dikin s method. During the period 1979 2003, there has been intense interest in the development of interior-point methods. These methods are related to classical least-square methods used in numerical analysis for making fits to data or fitting simpler functional forms to more complicated ones. Therefore interior research can tap into the vast literature of approximation theory. A theoretical breakthrough came in 1979: the Russian mathematician L. G. Khachian (based on the work of Shor, 1971 1972) discovered an ellipsoid algorithm whose running time in its worst case was significantly lower than that of the Simplex Algorithm in its worst case. Its iterates are not required to be feasible. Other theoretical results quickly followed, notably that of N. Karmarkar, who discovered an interior-point algorithm whose running-time performance in its worst case was significantly lower than that of Khachian s. This

69 was followed by the theoretical results of others that improved on the upper-bound estimates of the worst-case performance as the dimensions of the problems and the amount of input data increased indefinitely. The algorithm best suited for solving a particular problem or a special class of problems may not be the same algorithm best suited for solving any problem from the broad class of problems defined by Ax b, x 0, c T x = min. One criterion used for comparing algorithms is upper bounds on worst-case performance times as the dimensions of the problem grow indefinitely in size. This criterion turns out to be totally misleading for deciding which algorithm to use for practical problems because these theoretical upper-bound estimates are many many times greater than any experienced with practical problems. Attempts to characterize in a simple way the class (or classes) of practical problems from which one might be able to derive a theoretical explanation of the excellent performance times of some of the algorithms used in practice have, in general, failed. In special cases, such as the shortest-path problem, the performance of shortestpath algorithms for the entire class of shortest-path problems is comparable to that observed on actual problems. There has been progress proving that average performance on classes of randomly generated problems using a parametric variant of the Simplex Method resembles that obtained on practical problems, but no one claims these randomly generated problems are representative of the class of practical linear programs. Because the theoretical results can be totally misleading as to what algorithm to choose to solve a practical problem, an emperical approach is used. The linear programming profession has accumulated a large collection of test problems drawn from practical sources. These are used to compare the running times of various proposed algorithms. The general goal of these efforts is to find the algorithm that surpasses the performance of all other algorithms in the collection. For example, Karmarkar claimed (when he developed his method) that on very large problems his technique would be significantly faster. As of this writing, as far as the authors can ascertain, there appears to be no one algorithm that is a clear winner, i.e., that solves all (or almost all) of the test problems faster than all the other proposed methods. On problems with many bounding hyperplanes in the neighborhood of the optimum point, an interior method will probably do better than an exterior method. On problems with relatively few boundary planes (which is often the case in practice) an exterior method will be hard to beat. For this reason, it is likely that the commercial software of the future will be some sort of a hybrid because one does not know which kind of problem is being solved or because one wishes to obtain an extreme-point solution. Many specialized efficient codes have been proposed for solving structured linear programs such as network problems, staircase problems, block-angular systems, and multi-stage stochastic systems. Definition (Polynomial Time): Let the problem data size (length of the input data stream) be L, the total number of bits required to store the data of a linear program in m equations and n variables in the computer. An algorithm is said to have a polynomial worst-case running time if the algorithm s execu-

70 EARLY INTERIOR-POINT METHODS tion time (in, say, seconds) to find an optimal solution on the computer is less than some polynomial expression in L, m, and n. Otherwise the algorithm is said to be NP (nonpolynomial). Worst-Case Measures Can be Misleading. For example, given a linear program in m equations and n variables, it may be stated that a method requires less than O(n p m q ) iterations, where O(n p m q ) means some fixed constant times n p m q. If the constant for the worst-case bound were huge, say 10 100 (which may be larger than the number of all the electrons in the universe), then such a bound would be ridiculous. Implicit in such statements about a worst-case bound is the assumption that the fixed constant is small, say 10 or 100. Usually this assumption is valid, and it has become common practice to compare worst-case bounds of algorithms as if the fixed constants for each of the algorithms are the same. In general, given a linear program in m equations and n variables, projective methods require less than O(n) iterations. Path-following methods require less than O( n) iterations. Each of these also require O(n 3 ) arithmetic operations per iteration to solve a linear least-squares subproblem in order to find a steepest descent direction. However, with refinements (such as rank-one updates) it is possible to solve each least-squares problem in O(n 5/2 ) arithmetic operations instead of O(n 3 ). When the number of operations is multiplied by the bound on the number of iterations, we find that Karmarkar s projective method is bounded by O(n 7/2 ) arithmetic operations while path-following methods are bounded by O(n 3 ) arithmetic operations to obtain an optimal solution within a tolerance ɛ>0. The time required to carry out the arithmetic operations depend on L, the digits of input data. Thus the bound on the time to execute Karmarkar s algorithm is O(n 7/2 L). Because the number of arithmetic operations (and iterations) can depend critically on ɛ>0, the bound on the accuracy of the computed optimal solution, we will use the following definition of a polynomial-time algorithm. Definition (ɛ-optimal Polynomial Time): An algorithm to solve a linear program in m rows and n columns is said to have a polynomial worst-case running time (measured in seconds, say) if the time to execute it is less than some polynomial expression in L, m, n, q = log 10 ɛ,wherelisthe total number of bits required to store the problem s data in the computer, q is the number of significant decimal digits of accuracy of the optimal solution, and the tolerance ɛ is some measure of how much the calculated solution differs from the true optimal objective value, or an approximate feasible solution to a feasibility problem differs from the right-hand side of a true feasible solution. 3.1 VON NEUMANN S METHOD Von Neumann, in a discussion with George Dantzig in 1948, proposed the first interior algorithm for finding a feasible solution to a linear program with a convexity

http://www.springer.com/978-0-387-98613-5