VARIANTS OF THE SIMPLEX METHOD

Similar documents
4.1 The original problem and the optimal tableau

Some Advanced Topics in Linear Programming

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Linear Programming Problems

Section Notes 5. Review of Linear Programming. Applied Math / Engineering Sciences 121. Week of October 15, 2017

Linear Programming. Linear programming provides methods for allocating limited resources among competing activities in an optimal way.

Solutions for Operations Research Final Exam

Math 414 Lecture 30. The greedy algorithm provides the initial transportation matrix.

Section Notes 4. Duality, Sensitivity, and the Dual Simplex Algorithm. Applied Math / Engineering Sciences 121. Week of October 8, 2018

Linear Programming. Course review MS-E2140. v. 1.1

3. The Simplex algorithmn The Simplex algorithmn 3.1 Forms of linear programs

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini

The Ascendance of the Dual Simplex Method: A Geometric View

The Simplex Algorithm. Chapter 5. Decision Procedures. An Algorithmic Point of View. Revision 1.0

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs

4.1 Graphical solution of a linear program and standard form

Linear Optimization. Andongwisye John. November 17, Linkoping University. Andongwisye John (Linkoping University) November 17, / 25

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY

5.4 Pure Minimal Cost Flow

Read: H&L chapters 1-6

Civil Engineering Systems Analysis Lecture XV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics

Department of Mathematics Oleg Burdakov of 30 October Consider the following linear programming problem (LP):

Marginal and Sensitivity Analyses

AM 121: Intro to Optimization Models and Methods Fall 2017

Introduction. Linear because it requires linear functions. Programming as synonymous of planning.

The Simplex Algorithm

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Civil Engineering Systems Analysis Lecture XIV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics

NATCOR Convex Optimization Linear Programming 1

George B. Dantzig Mukund N. Thapa. Linear Programming. 1: Introduction. With 87 Illustrations. Springer

Solving Linear Programs Using the Simplex Method (Manual)

Linear programming II João Carlos Lourenço

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D.

What is linear programming (LP)? NATCOR Convex Optimization Linear Programming 1. Solving LP problems: The standard simplex method

Artificial Intelligence

5 The Theory of the Simplex Method

The Simplex Algorithm with a New. Primal and Dual Pivot Rule. Hsin-Der CHEN 3, Panos M. PARDALOS 3 and Michael A. SAUNDERS y. June 14, 1993.

3 INTEGER LINEAR PROGRAMMING

Graphs that have the feasible bases of a given linear

MATLAB Solution of Linear Programming Problems

Optimization of Design. Lecturer:Dung-An Wang Lecture 8

CASE STUDY. fourteen. Animating The Simplex Method. case study OVERVIEW. Application Overview and Model Development.

THE simplex algorithm [1] has been popularly used

CSC 8301 Design & Analysis of Algorithms: Linear Programming

Math Introduction to Operations Research

New Directions in Linear Programming

An iteration of the simplex method (a pivot )

16.410/413 Principles of Autonomy and Decision Making

Generalized Network Flow Programming

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Discrete Optimization. Lecture Notes 2

Math Models of OR: The Simplex Algorithm: Practical Considerations

Chapter 15 Introduction to Linear Programming

Linear programming and duality theory

Chapter II. Linear Programming

Linear Programming. Revised Simplex Method, Duality of LP problems and Sensitivity analysis

Fundamentals of Operations Research. Prof. G. Srinivasan. Department of Management Studies. Indian Institute of Technology Madras.

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Tuesday, April 10. The Network Simplex Method for Solving the Minimum Cost Flow Problem

Copyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch.

Easter Term OPTIMIZATION

Introductory Operations Research

maximize minimize b T y subject to A T y c, 0 apple y. The problem on the right is in standard form so we can take its dual to get the LP c T x

LP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008

ME 391Q Network Flow Programming

Programming, numerics and optimization

Linear Programming. Larry Blume. Cornell University & The Santa Fe Institute & IHS

A Generalized Model for Fuzzy Linear Programs with Trapezoidal Fuzzy Numbers

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

MA4254: Discrete Optimization. Defeng Sun. Department of Mathematics National University of Singapore Office: S Telephone:

COLUMN GENERATION IN LINEAR PROGRAMMING

Math 5593 Linear Programming Lecture Notes

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS

Integer Programming Theory

How to Solve a Standard Maximization Problem Using the Simplex Method and the Rowops Program

What s Linear Programming? Often your try is to maximize or minimize an objective within given constraints

Introduction to Operations Research

Duality. Primal program P: Maximize n. Dual program D: Minimize m. j=1 c jx j subject to n. j=1. i=1 b iy i subject to m. i=1

Part 1. The Review of Linear Programming The Revised Simplex Method

Pivot Rules for Linear Programming: A Survey on Recent Theoretical Developments

Linear Programming Motivation: The Diet Problem

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

Lecture notes on Transportation and Assignment Problem (BBE (H) QTM paper of Delhi University)

Lecture 9: Linear Programming

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Simulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Parallelizing the dual revised simplex method

Equilibrium Computation for Two-Player Games in Strategic and Extensive Form

x ji = s i, i N, (1.1)

Lecture Notes 2: The Simplex Algorithm

A Simplex Based Parametric Programming Method for the Large Linear Programming Problem

EARLY INTERIOR-POINT METHODS

6.854 Advanced Algorithms. Scribes: Jay Kumar Sundararajan. Duality

Recap, and outline of Lecture 18

An introduction to pplex and the Simplex Method

Transcription:

C H A P T E R 6 VARIANTS OF THE SIMPLEX METHOD By a variant of the Simplex Method (in this chapter) we mean an algorithm consisting of a sequence of pivot steps in the primal system using alternative rules for the selection of the pivot Historically these variants were developed to take advantage of a situation where an infeasible basic solution of the primal is available In other applications there often occurs a set of problems differing from one another only in their constant terms and cost factors In such cases, it is convenient to omit Phase I and to use the optimal basis of one problem as the initial basis for the next 61 INTRODUCTION Several methods have been proposed for varying the Simplex Algorithm to reduce the number of iterations This is especially needed for problems involving many equations in order to reduce the computation time It is also needed for problems involving a large number of variables n, for the number of iterations in practice appears to grow roughly proportional to n For example, instead of using the selection rule c s =min c j, one could select j = s such that introducing x s into the basic set gives the largest decrease in the value of z in the next basic solution This requires computing, for c j < 0, the largest in absolute value of c j θ j,whereθ j is determined so that if x j replaced x jr then the solution will remain feasible This rule is obviously not practical when using the revised Simplex Method with multipliers Even using the standard canonical form, considerably more computations would be required per iteration It is possible, however, in the nondegenerate case, to develop a modification of the canonical form in which the coefficient of the ith basic variable is allowed to be different from unity 173

174 VARIANTS OF THE SIMPLEX METHOD in the ith equation but b i = 1 In this form the selection of s by the steepest descent criterion would require a little more effort; moreover (by means of a special device), no more effort than that for the standard Simplex Algorithm would be required to maintain the tableau in proper modified form from iteration to iteration (see Section 62 for details) The simplest variant occurs when the new problem differs from the original in the cost coefficients alone In this case, the cost coefficients are replaced by the new ones, and Phase II of the Revised Simplex Method is applied Another possibility is to use the parametric objective method to be described in Section 64 An important variant occurs when the new problem differs from the original in the constant terms only In this case the optimal basis of the first problem will still price out dual feasible, ie, c j 0, for the second, but the associated primal solution may not be feasible For this situation, we could use the Dual-Simplex Algorithm, which is the variant of the standard Primal-Simplex Algorithm, to be discussed in Section 63, or the paramteric right-hand-side method to be described in Section 64, or the Primal-Dual method of Section 66 However, when the problems differ by more than either the constant terms or the cost coefficient terms, the old basis may be neither primal feasible nor dual feasible When neither the basic solution nor the dual solution generated by its simplex multipliers remains feasible, the corresponding algorithm is called composite The Self-Dual parametric algorithm discussed in Section 65 is an example of such a composite algorithm Correspondence of Primal and Dual Bases In 1954, Lemke discovered a certain correspondence between the bases of the primal and dual systems that made it possible to interpret the Simplex Algorithm as applied to the dual as a sequence of basis changes in the primal; this interpretation is called the Dual-Simplex Algorithm From a computational point of view, the Dual- Simplex Algorithm is advantageous because the size of the basis being manipulated in the computer is m m instead of n n In this case, however, the associated basic solutions of the primal are not feasible, but the simplex multipliers continue to price out optimal (hence, yield a basic feasible solution to the dual) It is good to understand the details of this correspondence, for it provides a means of easily dualizing a problem without transposing the constraint matrix Consider the standard form Minimize c T x = z subject to Ax = b, A : m n, x 0, (61) and the dual of the standard form: Maximize b T π = v subject to A T π c, A : m n, (62)

61 INTRODUCTION 175 PRIMAL-DUAL CORRESPONDENCES Primal Dual ( ) B Basis B B T 0 = N T I Basic variables x B π, y N = c N Nonbasic variables x N y B = c B Feasibility condition Ax = b, x 0 c 0 Table 6-1: Primal-Dual Correspondences where π i is unrestricted in sign We assume A is full rank Adding slack variables y 0 to the dual problem (62) we get: Maximize b T π = v subject to A T π + Iy = c, A : m n, y 0 (63) Clearly the dual has n basic variables and m nonbasic variables The variables π areunrestrictedinsignand,whena is full rank, always constitute m out of the n basic variables of the dual The basis for the primal is denoted by B and the nonbasic columns are denoted by N Note that y = c A T π = c 0 when the dual is feasible and that c B =0and c N 0 when the primal is optimal The basic and nonbasic columns for the dual are denoted by B, ann n matrix, and N, an n m matrix ( ) ( B T 0 Im B = N T, N = I n m 0 ), (64) where I n m is an (n m)-dimensional identity matrix and I m is an m-dimensional identity matrix Thus, (π, y N ) as basic variables and y B as nonbasic variables constitute a basic solution to the dual if and only if y B =0 Exercise 61 Show that the determinant of B has the same value as that of B Also show that if B 1 exists, then B 1 exists It is now clear that there is a correspondence between primal and dual bases These correspondences are shown in Table 6-1 With these correspondences in mind, we shall discuss variations of the Simplex Method Exercise 62 How is the concept of complementary primal and dual variables related to the correspondence of primal and dual bases? Definition (Dual Degeneracy): We have already defined degeneracy and nondegeneracy with respect to the primal A basis is said to be dual degenerate

176 VARIANTS OF THE SIMPLEX METHOD if one or more of the c j corresponding to nonbasic variables x j are zero and to be dual nondegenerate otherwise 62 MAX IMPROVEMENT PER ITERATION In this section we present an alternative canonical form for efficiently determining the incoming column that yields the maximum improvement per iteration It has been observed on many test problems that this often leads to fewer iterations We will assume, to simplify the discussion, that all basic feasible solutions that are generated are nondegenerate Assume x j 0forj =1,,n and x j for j =1,,m are basic variables, in the standard canonical form: z + c m+1 x m+1 + + c j x j + + c n x n = z 0 x 1 + ā 1,m+1 x m+1 + + ā 1j x j + + ā 1n x n = b 1 x p + ā p,m+1 x m+1 + + ā pj x j + + ā pn x n = b p x m +ā m,m+1 x m+1 + +ā mj x j + +ā mn x n = b m (65) where ā ij, c j, b i,and z 0 are constants and we assume that b i > 0fori =1,,mIf there are two or more c j < 0, our problem is to find which j = s among them has the property that putting column s into the basis and driving some column j = r out of the basis gives the greatest decrease in z while preserving feasibility Using the standard canonical form (65), the main work is thatof performing ratiotests for the several columns j such that c j < 0, plus the update work of pivoting, which takes mn operations where each operation consists of one addition (or subtraction) plus one multiplication To do this more efficiently, consider the alternative canonical form: z + c m+1 x m+1 + + c n x n = z 0 α 11 x 1 + α 1,m+1 x m+1 + + α 1n x n = 1 α pp x p + α p,m+1 x m+1 + + α pn x n = 1 α mm x m + α m,m+1 x m+1 + + α mn x n = 1 (66) formed by rescaling the rows by dividing them by b i > 0 for all i =1,,m In this format we scan column j such that c j < 0 and find for each such j the row r j =argmax α ij

http://wwwspringercom/978-0-387-98613-5