The Affine Scaling Method
|
|
- Michael Gibson
- 5 years ago
- Views:
Transcription
1 MA33 Linear Programming W. J. Martin October 9, 8 The Affine Scaling Method Overview Given a linear programming problem in equality form with full rank constraint matrix and a strictly positive feasible solution x, we transform the problem to a new one with feasible solution = [,,,. The original problem and scaled problem Original: Scaled: max c x max ĉ x s. t. Ax = b s. t. Âx = b x x are related as follows: X is the diagonal matrix with the entries of x down the diagonal; Â = AX ; and ĉ = X c. In the scaled problem, we project the gradient ĉ onto the subspace Âx = to obtain a search direction d; find the largest scalar β such that + βd is still non-negative (the other feasibility conditions are already guaranteed by choice of d); obtain new (scaled) solution ˆx = + rβd where < r < is a global constant; update our solution x to the next solution x = X ˆx. This process is repeated, as needed, until we are sufficiently close to an optimal solution (e.g., as one may check by solving for a dual vector using complementary slackness conditions).
2 One iteration of the affine scaling method We assume we have an m n matrix A with full row rank, an m-vector b and an n-vector c. (If A does not have full row rank, first row reduce, and if Ax = b is feasible, then replace A by a smaller matrix.) We also have a fixed step size r, < r <. Given a strictly positive feasible solution x to the problem (so Ax = b), we obtain a better strictly positive feasible solution x as follows: Step : (Scale) Let  = AX and ĉ = X c where X is the diagonal matrix with entries from x down the diagonal. (Now we have a scaled problem with as a feasible solution.) Step : (Find projection) Compute the projection matrix onto the nullspace of Â: P = I  ( )  Step 3: (Find search direction) The search direction (in the scaled problem) is then d = P ĉ. { } Step 4: (Compute ratios) Let β = min d j : d j <. (If this set is empty, the problem is unbounded.) Step 5: (Scaled update) The improved solution in the scaled problem is ˆx = + rβd where r is the prescribed step size ( < r < ). Step 6: (Next iterate) Now scale back to the original problem and take x = X ˆx. Note that the above algorithm brings us from x to x. To obtain further improved solutions, we iterate this with x replaced by the current solution. The general step is described with only a slight change of notation and is given at the end of these notes, just before the exercises.
3 Some explanations In Step, we simply write down the data (Â, ĉ) for the scaled problem. right-hand side vector does not change. But that s okay because Note that the  = (AX ) = A(X ) = Ax = b so is indeed a feasible solution to the scaled problem and it is far away from the boundaries x j =. Moreover, X is a diagonal matrix, so it is symmetric: (X ) = X. Thus ĉ = (X c) = (c (X ) ) = c (X ) = c x. More generally, the objective value of any solution ˆx to the scaled problem is equal to the objective value of the corresponding solution x = X ˆx in the original problem. In Step, we do the most work. We need to project the gradient ĉ onto the subspace {x : Âx = } in order to have a feasible search direction. As we worked out in class, the projection matrix P does just the trick. In Step 3, we take d = P ĉ so that Âd = and, for any scalar β, the vector ˆx = + βd satisfies ˆx = Â( + βd) =  + βâd =  = b as established above. So, no matter how far we travel in this direction, the equality constraints will still be satisfied. From this, we know in Step 4 that any vector of the form ˆx = +βd satisfies the equality constraints. So in order to stay feasible, we simply need +βd. This requires +βd j for all j. Since we are going to take β positive (to improve the objective value), the only coordinates j that concern us are those with d j <, and each of these gives us an upper limit β /d j. In Step 5, we finally use our global step size r. We know that +βd and is otherwise feasible. But we need a strictly positive feasible solution. So we only move a fraction of the maximum possible distance. For example, with r =.9, we only go 9% of the way to the wall. Thus our update is + r(βd). The transformed problem has now served its purpose of pushing the search direction away from the boundaries. We discard it and transform the solution ˆx back to the original LP to get the next iterate x. As explained in Step, this is simply achieved through scaling by X. 3
4 Example : Consider the problem max 3 x + x s.t. 5 x + x + x 3 = x + x + x 4 = x, x, x 3, x 4 with initial feasible solution x = (,,, ). So our current objective value is ζ = We have A = [ 5 Step : Scale the problem: /5 X = /5 /5 Step : Compute P : [, b =,  = AX =  = [ 5, c = 3, x = [ 4/5 /5 /5 4/5 ) P = I (  Â. [, ( ) = 5 63 We eliminate fractions where possible to write 5 ) [  ( 5 = so that  ( )  = Let s call this last matrix Q. Then we have P I Q = 63 [ = 5 63 = /5 /5 /5., ĉ = X c = /5 /5.
5 Step 3: Find search direction d: d = P ĉ = which simplifies to d = ( 3,, 7, ). 63 3/5 /5 = Step 4: Ratios: So β = 9. { } { 63 β = min : d j < = min d j 7, 63 } = 9. Step 5: Scaled update (using step size r = /): 4 ˆx = + rβd = using rβ 63 = 9 63 = 4. Step 6: Next iterate: Now we go back to the original problem using X : /5 x = X ˆx = /5 / /4 /4 7/4 5/4 = 4 = 7 with objective value ζ = 5/4. As you can see, the numbers will get worse after this, so we use a computer if we want to go further Example : Consider the problem max x x x + x = 4 x, x with initial feasible solution x = (, ) and step size r = /. We have A = [, b = [ 4 [, c =, x = [. 5
6 Step : Scale the problem: X = [ Step : Compute P = I Â (ÂÂ ) Â: ÂÂ = [ [, Â = AX = [ [, ĉ = X c = = [ 8, P = [ P = Step 3: Find search direction d: d = P ĉ = [ [ [ [ / / / / = [. [ /8 [ = [ 3 3 [ = 3/ 3/ [ Step 4: Ratios: { } β = min : d j < = d j 3. Step 5: Scaled update: Step 6: Next iterate: ˆx = + rβd = [ x = X ˆx = [ + 3 [ [ 3/ / 3 3 = = [ 3/ [ 3/ / In this iteration, our objective value has improved from to +/. Pictorial Example The advantage of interior point methods is that they tend to avoid the combinatorial complexity of basic feasible solutions. One can imagine a high-dimensional polyhedron having large clusters of nearby vertices all of which are suboptimal. The simplex method can only move from one vertex to a neighbor, possibly contributing to inefficiency.. 6
7 To illustrate how an interior point approach can smooth out all these corners, let s look at a simple two-dimensional example with three nearby basic feasible solutions far away from optimality: maximize x subject to x x x x 9 x x 8 x 8 x, x We will skip the conversion to equality form. After introducing slack variables x 3,..., x 6, the initial strictly feasible solution (4, ) gives us x = (4,, 5,,, 5 ). The objective direction c is horizontal, pointing directly to the right. But the optimal solution is at (4, 8). Here we give the first eight iterates x,..., x 8 of the affine scaling method with two choices of step size, r = / and r = 3/4. top red curve r = bottom green curve r = 3 4 x c 7
8 Initially, we showed how to move from initial solution x to a better solution x. generic iteration looks like this: A Step : (Scale) Let  = AXk and ĉ = X k c where X k is the diagonal matrix with entries from the most recent solution x k down the diagonal. (Now we have a scaled problem with as a feasible solution.) Step : (Find projection) Compute the projection matrix onto the nullspace of Â: P = I  ( )  Step 3: (Find search direction) The search direction (in the scaled problem) is then d = P ĉ. { } Step 4: (Compute ratios) Let β = min d j : d j <. (If this set is empty, the problem is unbounded.) Step 5: (Scaled update) The improved solution in the scaled problem is ˆx = + rβd where r is the prescribed step size ( < r < ). Step 6: (Next iterate) Now scale back to the original problem to obtain the next iterate x k+ = X kˆx. Exercises.) Consider the linear programming problem max x x subject to x + x, x, x. (i) Starting with x = (/, /4), apply two iterations of the affine scaling method using r = /. (First, convert to equality form.) For each iteration, give the constraint matrix  and objective vector ĉ for the scaled problem; the projection matrix P and search direction d for this iteration; the ratio computation from Step 4; 8
9 the next iterate, both in scaled form ˆx and as a solution x k+ to the original problem above. (So you will be finding x and x.) (ii) On a sheet of graph paper, make a careful (and large!) drawing of the feasible region in R. For x and each of the next two iterates, x, and x, plot both the gradient of the objective function (namely c = [ ) at that point as well as the scaled step direction x k+ x k..) Repeat the steps of Exercise (again with r = /) for the linear programming problem max x + x subject to x, x. starting with x = (3/4, /4) (a column vector). (You may wish to use MAPLE to check your computations and to follow the trajectory further.) 3.) Perform two iterations of the affine scaling method for the problem maximize x + x + x 3 subject to x x = x + x + x 3 = 4 x, x, x 3 with step size r = / and initial solution x = (,, ). 4.) Perform two iterations of the affine scaling method for the problem max x subject to x + x = 7, x, x with step size r = 3 4 and initial solution x = (3, 4). 5.) Perform two iterations of the affine scaling method for the problem max x x + 3x 3 subject to x + x + x 3 = 5 (all variables nonnegative) with step size r = and initial solution x = (,, ). 6.) Perform two iterations of the affine scaling method for the problem max c x s.t. Ax = b, x with step size r = / and initial solution x = (5, 3, ) where [ [ 5 A =, b =, c = (,, ). 7 7.) Perform two iterations of the affine scaling method for the problem maximize x x 3 subject to x + 4x 4x 3 = x + x + x 3 = 8 x, x, x 3 with step size r = / and initial solution x = (,, ). 9
10 8.) Perform two iterations of the affine scaling method for the problem max x + x s.t. x + 5x = 3 x, x with step size r =.75 and initial solution x = (,.). 9.) Perform two iterations of the affine scaling method for the problem maximize x x 3 subject to x x = x + x 3 = x, x, x 3 with step size r = / and initial solution x = (,, )..) Perform two iterations of the affine scaling method for the problem min x x s.t. 3x + 5x = 3, x, x with step size r = / and initial solution x = (, )..) Perform two iterations of the affine scaling method for the problem maximize x + x + x 3 subject to x x = x + x + x 3 = 4 x, x, x 3 with step size r = / and initial solution x = (,, )..) More generically, consider the n-dimensional optimization problem with one equality constraint n j= x j =, objective function max c x, step size r = /, and initial solution x with positive entries summing to one. For convenience, also assume that n l= c l(x l ) = x and that the variables are ordered in such a way that c c c n. Compare the next iterate x produced by the affine scaling method against the corresponding unscaled update x + β c where c is the projection of the objective gradient c onto the affine subspace containing the feasible region and β is computed so as to prevent any entry of x + β c from going negative.
LECTURE 6: INTERIOR POINT METHOD. 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm
LECTURE 6: INTERIOR POINT METHOD 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm Motivation Simplex method works well in general, but suffers from exponential-time
More informationCS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm Instructor: Shaddin Dughmi Algorithms for Convex Optimization We will look at 2 algorithms in detail: Simplex and Ellipsoid.
More informationMath 5593 Linear Programming Lecture Notes
Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................
More information16.410/413 Principles of Autonomy and Decision Making
16.410/413 Principles of Autonomy and Decision Making Lecture 17: The Simplex Method Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology November 10, 2010 Frazzoli (MIT)
More informationSome Advanced Topics in Linear Programming
Some Advanced Topics in Linear Programming Matthew J. Saltzman July 2, 995 Connections with Algebra and Geometry In this section, we will explore how some of the ideas in linear programming, duality theory,
More information4 LINEAR PROGRAMMING (LP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1
4 LINEAR PROGRAMMING (LP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1 Mathematical programming (optimization) problem: min f (x) s.t. x X R n set of feasible solutions with linear objective function
More information4.1 Graphical solution of a linear program and standard form
4.1 Graphical solution of a linear program and standard form Consider the problem min c T x Ax b x where x = ( x1 x ) ( 16, c = 5 ), b = 4 5 9, A = 1 7 1 5 1. Solve the problem graphically and determine
More informationAMS : Combinatorial Optimization Homework Problems - Week V
AMS 553.766: Combinatorial Optimization Homework Problems - Week V For the following problems, A R m n will be m n matrices, and b R m. An affine subspace is the set of solutions to a a system of linear
More informationSection Notes 5. Review of Linear Programming. Applied Math / Engineering Sciences 121. Week of October 15, 2017
Section Notes 5 Review of Linear Programming Applied Math / Engineering Sciences 121 Week of October 15, 2017 The following list of topics is an overview of the material that was covered in the lectures
More information3 INTEGER LINEAR PROGRAMMING
3 INTEGER LINEAR PROGRAMMING PROBLEM DEFINITION Integer linear programming problem (ILP) of the decision variables x 1,..,x n : (ILP) subject to minimize c x j j n j= 1 a ij x j x j 0 x j integer n j=
More informationMathematical and Algorithmic Foundations Linear Programming and Matchings
Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis
More informationRead: H&L chapters 1-6
Viterbi School of Engineering Daniel J. Epstein Department of Industrial and Systems Engineering ISE 330: Introduction to Operations Research Fall 2006 (Oct 16): Midterm Review http://www-scf.usc.edu/~ise330
More informationCS 473: Algorithms. Ruta Mehta. Spring University of Illinois, Urbana-Champaign. Ruta (UIUC) CS473 1 Spring / 29
CS 473: Algorithms Ruta Mehta University of Illinois, Urbana-Champaign Spring 2018 Ruta (UIUC) CS473 1 Spring 2018 1 / 29 CS 473: Algorithms, Spring 2018 Simplex and LP Duality Lecture 19 March 29, 2018
More informationLinear Programming. Linear programming provides methods for allocating limited resources among competing activities in an optimal way.
University of Southern California Viterbi School of Engineering Daniel J. Epstein Department of Industrial and Systems Engineering ISE 330: Introduction to Operations Research - Deterministic Models Fall
More informationAdvanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material
More informationThe Simplex Algorithm
The Simplex Algorithm Uri Feige November 2011 1 The simplex algorithm The simplex algorithm was designed by Danzig in 1947. This write-up presents the main ideas involved. It is a slight update (mostly
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More informationMath 414 Lecture 30. The greedy algorithm provides the initial transportation matrix.
Math Lecture The greedy algorithm provides the initial transportation matrix. matrix P P Demand W ª «2 ª2 «W ª «W ª «ª «ª «Supply The circled x ij s are the initial basic variables. Erase all other values
More informationmaximize c, x subject to Ax b,
Lecture 8 Linear programming is about problems of the form maximize c, x subject to Ax b, where A R m n, x R n, c R n, and b R m, and the inequality sign means inequality in each row. The feasible set
More informationLecture 4: Linear Programming
COMP36111: Advanced Algorithms I Lecture 4: Linear Programming Ian Pratt-Hartmann Room KB2.38: email: ipratt@cs.man.ac.uk 2017 18 Outline The Linear Programming Problem Geometrical analysis The Simplex
More informationOptimization of Design. Lecturer:Dung-An Wang Lecture 8
Optimization of Design Lecturer:Dung-An Wang Lecture 8 Lecture outline Reading: Ch8 of text Today s lecture 2 8.1 LINEAR FUNCTIONS Cost Function Constraints 3 8.2 The standard LP problem Only equality
More informationLinear programming and duality theory
Linear programming and duality theory Complements of Operations Research Giovanni Righini Linear Programming (LP) A linear program is defined by linear constraints, a linear objective function. Its variables
More informationLinear Programming Terminology
Linear Programming Terminology The carpenter problem is an example of a linear program. T and B (the number of tables and bookcases to produce weekly) are decision variables. The profit function is an
More informationThe Simplex Algorithm for LP, and an Open Problem
The Simplex Algorithm for LP, and an Open Problem Linear Programming: General Formulation Inputs: real-valued m x n matrix A, and vectors c in R n and b in R m Output: n-dimensional vector x There is one
More informationLinear Programming Problems
Linear Programming Problems Two common formulations of linear programming (LP) problems are: min Subject to: 1,,, 1,2,,;, max Subject to: 1,,, 1,2,,;, Linear Programming Problems The standard LP problem
More informationLecture 6: Faces, Facets
IE 511: Integer Programming, Spring 2019 31 Jan, 2019 Lecturer: Karthik Chandrasekaran Lecture 6: Faces, Facets Scribe: Setareh Taki Disclaimer: These notes have not been subjected to the usual scrutiny
More informationInteger Programming Theory
Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x
More informationEllipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case
Ellipsoid Algorithm 15-853:Algorithms in the Real World Linear and Integer Programming II Ellipsoid algorithm Interior point methods First polynomial-time algorithm for linear programming (Khachian 79)
More informationPOLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if
POLYHEDRAL GEOMETRY Mathematical Programming Niels Lauritzen 7.9.2007 Convex functions and sets Recall that a subset C R n is convex if {λx + (1 λ)y 0 λ 1} C for every x, y C and 0 λ 1. A function f :
More informationSection Notes 4. Duality, Sensitivity, and the Dual Simplex Algorithm. Applied Math / Engineering Sciences 121. Week of October 8, 2018
Section Notes 4 Duality, Sensitivity, and the Dual Simplex Algorithm Applied Math / Engineering Sciences 121 Week of October 8, 2018 Goals for the week understand the relationship between primal and dual
More informationLinear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage).
Linear Programming Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory: Feasible Set, Vertices, Existence of Solutions. Equivalent formulations. Outline
More informationLinear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.
Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.
More informationPrograms. Introduction
16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen
More informationLinear Programming and its Applications
Linear Programming and its Applications Outline for Today What is linear programming (LP)? Examples Formal definition Geometric intuition Why is LP useful? A first look at LP algorithms Duality Linear
More information16.410/413 Principles of Autonomy and Decision Making
16.410/413 Principles of Autonomy and Decision Making Lecture 16: Mathematical Programming I Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology November 8, 2010 E. Frazzoli
More informationFinite Math Linear Programming 1 May / 7
Linear Programming Finite Math 1 May 2017 Finite Math Linear Programming 1 May 2017 1 / 7 General Description of Linear Programming Finite Math Linear Programming 1 May 2017 2 / 7 General Description of
More informationMath 273a: Optimization Linear programming
Math 273a: Optimization Linear programming Instructor: Wotao Yin Department of Mathematics, UCLA Fall 2015 some material taken from the textbook Chong-Zak, 4th Ed. History The word programming used traditionally
More informationLecture 16 October 23, 2014
CS 224: Advanced Algorithms Fall 2014 Prof. Jelani Nelson Lecture 16 October 23, 2014 Scribe: Colin Lu 1 Overview In the last lecture we explored the simplex algorithm for solving linear programs. While
More informationArtificial Intelligence
Artificial Intelligence Combinatorial Optimization G. Guérard Department of Nouvelles Energies Ecole Supérieur d Ingénieurs Léonard de Vinci Lecture 1 GG A.I. 1/34 Outline 1 Motivation 2 Geometric resolution
More informationIE 5531: Engineering Optimization I
IE 5531: Engineering Optimization I Lecture 3: Linear Programming, Continued Prof. John Gunnar Carlsson September 15, 2010 Prof. John Gunnar Carlsson IE 5531: Engineering Optimization I September 15, 2010
More informationLECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach
LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction
More informationLinear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?
Linear and Integer Programming 15-853:Algorithms in the Real World Linear and Integer Programming I Introduction Geometric Interpretation Simplex Method Linear or Integer programming maximize z = c T x
More informationLecture Notes 2: The Simplex Algorithm
Algorithmic Methods 25/10/2010 Lecture Notes 2: The Simplex Algorithm Professor: Yossi Azar Scribe:Kiril Solovey 1 Introduction In this lecture we will present the Simplex algorithm, finish some unresolved
More informationVARIANTS OF THE SIMPLEX METHOD
C H A P T E R 6 VARIANTS OF THE SIMPLEX METHOD By a variant of the Simplex Method (in this chapter) we mean an algorithm consisting of a sequence of pivot steps in the primal system using alternative rules
More informationLecture 2 - Introduction to Polytopes
Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.
More informationDM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini
DM545 Linear and Integer Programming Lecture 2 The Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. 2. 3. 4. Standard Form Basic Feasible Solutions
More informationConvex Optimization CMU-10725
Convex Optimization CMU-10725 Ellipsoid Methods Barnabás Póczos & Ryan Tibshirani Outline Linear programs Simplex algorithm Running time: Polynomial or Exponential? Cutting planes & Ellipsoid methods for
More informationExtensions of Semidefinite Coordinate Direction Algorithm. for Detecting Necessary Constraints to Unbounded Regions
Extensions of Semidefinite Coordinate Direction Algorithm for Detecting Necessary Constraints to Unbounded Regions Susan Perrone Department of Mathematics and Statistics Northern Arizona University, Flagstaff,
More informationLec13p1, ORF363/COS323
Lec13 Page 1 Lec13p1, ORF363/COS323 This lecture: Semidefinite programming (SDP) Definition and basic properties Review of positive semidefinite matrices SDP duality SDP relaxations for nonconvex optimization
More informationDesign and Analysis of Algorithms (V)
Design and Analysis of Algorithms (V) An Introduction to Linear Programming Guoqiang Li School of Software, Shanghai Jiao Tong University Homework Assignment 2 is announced! (deadline Apr. 10) Linear Programming
More informationOutline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014
5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38
More informationConic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding
Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding B. O Donoghue E. Chu N. Parikh S. Boyd Convex Optimization and Beyond, Edinburgh, 11/6/2104 1 Outline Cone programming Homogeneous
More informationCOLUMN GENERATION IN LINEAR PROGRAMMING
COLUMN GENERATION IN LINEAR PROGRAMMING EXAMPLE: THE CUTTING STOCK PROBLEM A certain material (e.g. lumber) is stocked in lengths of 9, 4, and 6 feet, with respective costs of $5, $9, and $. An order for
More informationLinear Programming. them such that they
Linear Programming l Another "Sledgehammer" in our toolkit l Many problems fit into the Linear Programming approach l These are optimization tasks where both the constraints and the objective are linear
More informationSolutions for Operations Research Final Exam
Solutions for Operations Research Final Exam. (a) The buffer stock is B = i a i = a + a + a + a + a + a 6 + a 7 = + + + + + + =. And the transportation tableau corresponding to the transshipment problem
More informationChapter II. Linear Programming
1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION
More information60 2 Convex sets. {x a T x b} {x ã T x b}
60 2 Convex sets Exercises Definition of convexity 21 Let C R n be a convex set, with x 1,, x k C, and let θ 1,, θ k R satisfy θ i 0, θ 1 + + θ k = 1 Show that θ 1x 1 + + θ k x k C (The definition of convexity
More informationLinear Programming 1
Linear Programming 1 Fei Li March 5, 2012 1 With references of Algorithms by S. Dasgupta, C. H. Papadimitriou, and U. V. Vazirani. Many of the problems for which we want algorithms are optimization tasks.
More informationAlgebra Reviews & LP Graphic Solutions
Algebra Reviews & LP Graphic Solutions Given Constraints to Draw Straight Lines and Identify Feasible Region Draw Straight Lines for Each Constraint: From Equ(1), Set X = 0, Y = 3, a(0, 3); Set Y = 0,
More informationThe Simplex Algorithm
The Simplex Algorithm April 25, 2005 We seek x 1,..., x n 0 which mini- Problem. mizes C(x 1,..., x n ) = c 1 x 1 + + c n x n, subject to the constraint Ax b, where A is m n, b = m 1. Through the introduction
More informationLecture 3. Corner Polyhedron, Intersection Cuts, Maximal Lattice-Free Convex Sets. Tepper School of Business Carnegie Mellon University, Pittsburgh
Lecture 3 Corner Polyhedron, Intersection Cuts, Maximal Lattice-Free Convex Sets Gérard Cornuéjols Tepper School of Business Carnegie Mellon University, Pittsburgh January 2016 Mixed Integer Linear Programming
More informationLecture 9: Linear Programming
Lecture 9: Linear Programming A common optimization problem involves finding the maximum of a linear function of N variables N Z = a i x i i= 1 (the objective function ) where the x i are all non-negative
More informationCopyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch.
Iterative Improvement Algorithm design technique for solving optimization problems Start with a feasible solution Repeat the following step until no improvement can be found: change the current feasible
More information3. The Simplex algorithmn The Simplex algorithmn 3.1 Forms of linear programs
11 3.1 Forms of linear programs... 12 3.2 Basic feasible solutions... 13 3.3 The geometry of linear programs... 14 3.4 Local search among basic feasible solutions... 15 3.5 Organization in tableaus...
More informationLP Graphic Solution & Solver for LP
LP Graphic Solution & Excel@ Solver for LP 5X + 3Y = 0 (3) Draw the objective function line on the graph: Any (including the optimal) LP solution has to be part of
More informationCivil Engineering Systems Analysis Lecture XV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics
Civil Engineering Systems Analysis Lecture XV Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics Today s Learning Objectives Sensitivity Analysis Dual Simplex Method 2
More informationMATH 890 HOMEWORK 2 DAVID MEREDITH
MATH 890 HOMEWORK 2 DAVID MEREDITH (1) Suppose P and Q are polyhedra. Then P Q is a polyhedron. Moreover if P and Q are polytopes then P Q is a polytope. The facets of P Q are either F Q where F is a facet
More informationIntroduction. Linear because it requires linear functions. Programming as synonymous of planning.
LINEAR PROGRAMMING Introduction Development of linear programming was among the most important scientific advances of mid-20th cent. Most common type of applications: allocate limited resources to competing
More informationDiscrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity
Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity Marc Uetz University of Twente m.uetz@utwente.nl Lecture 5: sheet 1 / 26 Marc Uetz Discrete Optimization Outline 1 Min-Cost Flows
More information6.854 Advanced Algorithms. Scribes: Jay Kumar Sundararajan. Duality
6.854 Advanced Algorithms Scribes: Jay Kumar Sundararajan Lecturer: David Karger Duality This lecture covers weak and strong duality, and also explains the rules for finding the dual of a linear program,
More informationMathematical Programming and Research Methods (Part II)
Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types
More information5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY
5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY 5.1 DUALITY Associated with every linear programming problem (the primal) is another linear programming problem called its dual. If the primal involves
More informationCMPSCI611: The Simplex Algorithm Lecture 24
CMPSCI611: The Simplex Algorithm Lecture 24 Let s first review the general situation for linear programming problems. Our problem in standard form is to choose a vector x R n, such that x 0 and Ax = b,
More informationAdvanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 2 Review Dr. Ted Ralphs IE316 Quiz 2 Review 1 Reading for The Quiz Material covered in detail in lecture Bertsimas 4.1-4.5, 4.8, 5.1-5.5, 6.1-6.3 Material
More informationLinear Programming. Larry Blume. Cornell University & The Santa Fe Institute & IHS
Linear Programming Larry Blume Cornell University & The Santa Fe Institute & IHS Linear Programs The general linear program is a constrained optimization problem where objectives and constraints are all
More information1. Lecture notes on bipartite matching February 4th,
1. Lecture notes on bipartite matching February 4th, 2015 6 1.1.1 Hall s Theorem Hall s theorem gives a necessary and sufficient condition for a bipartite graph to have a matching which saturates (or matches)
More informationAM 121: Intro to Optimization Models and Methods Fall 2017
AM 121: Intro to Optimization Models and Methods Fall 2017 Lecture 10: Dual Simplex Yiling Chen SEAS Lesson Plan Interpret primal simplex in terms of pivots on the corresponding dual tableau Dictionaries
More informationChapter 15 Introduction to Linear Programming
Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of
More informationGraphs that have the feasible bases of a given linear
Algorithmic Operations Research Vol.1 (2006) 46 51 Simplex Adjacency Graphs in Linear Optimization Gerard Sierksma and Gert A. Tijssen University of Groningen, Faculty of Economics, P.O. Box 800, 9700
More informationIntroduction to Mathematical Programming IE406. Lecture 4. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 4 Dr. Ted Ralphs IE406 Lecture 4 1 Reading for This Lecture Bertsimas 2.2-2.4 IE406 Lecture 4 2 The Two Crude Petroleum Example Revisited Recall the
More informationMulti-level Partition of Unity Implicits
Multi-level Partition of Unity Implicits Diego Salume October 23 rd, 2013 Author: Ohtake, et.al. Overview Goal: Use multi-level partition of unity (MPU) implicit surface to construct surface models. 3
More informationAMATH 383 Lecture Notes Linear Programming
AMATH 8 Lecture Notes Linear Programming Jakob Kotas (jkotas@uw.edu) University of Washington February 4, 014 Based on lecture notes for IND E 51 by Zelda Zabinsky, available from http://courses.washington.edu/inde51/notesindex.htm.
More informationName: THE SIMPLEX METHOD: STANDARD MAXIMIZATION PROBLEMS
Name: THE SIMPLEX METHOD: STANDARD MAXIMIZATION PROBLEMS A linear programming problem consists of a linear objective function to be maximized or minimized subject to certain constraints in the form of
More informationSubmodularity Reading Group. Matroid Polytopes, Polymatroid. M. Pawan Kumar
Submodularity Reading Group Matroid Polytopes, Polymatroid M. Pawan Kumar http://www.robots.ox.ac.uk/~oval/ Outline Linear Programming Matroid Polytopes Polymatroid Polyhedron Ax b A : m x n matrix b:
More informationLinear Programming Motivation: The Diet Problem
Agenda We ve done Greedy Method Divide and Conquer Dynamic Programming Network Flows & Applications NP-completeness Now Linear Programming and the Simplex Method Hung Q. Ngo (SUNY at Buffalo) CSE 531 1
More information5. Consider the tree solution for a minimum cost network flow problem shown in Figure 5.
FIGURE 1. Data for a network flow problem. As usual the numbers above the nodes are supplies (negative values represent demands) and numbers shown above the arcs are unit shipping costs. The darkened arcs
More informationLecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.
Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i
More informationThe Ascendance of the Dual Simplex Method: A Geometric View
The Ascendance of the Dual Simplex Method: A Geometric View Robert Fourer 4er@ampl.com AMPL Optimization Inc. www.ampl.com +1 773-336-AMPL U.S.-Mexico Workshop on Optimization and Its Applications Huatulco
More informationLP Geometry: outline. A general LP. minimize x c T x s.t. a T i. x b i, i 2 M 1 a T i x = b i, i 2 M 3 x j 0, j 2 N 1. where
LP Geometry: outline I Polyhedra I Extreme points, vertices, basic feasible solutions I Degeneracy I Existence of extreme points I Optimality of extreme points IOE 610: LP II, Fall 2013 Geometry of Linear
More informationNATCOR Convex Optimization Linear Programming 1
NATCOR Convex Optimization Linear Programming 1 Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk 5 June 2018 What is linear programming (LP)? The most important model used in
More information4 Linear Programming (LP) E. Amaldi -- Foundations of Operations Research -- Politecnico di Milano 1
4 Linear Programming (LP) E. Amaldi -- Foundations of Operations Research -- Politecnico di Milano 1 Definition: A Linear Programming (LP) problem is an optimization problem: where min f () s.t. X n the
More informationPolytopes Course Notes
Polytopes Course Notes Carl W. Lee Department of Mathematics University of Kentucky Lexington, KY 40506 lee@ms.uky.edu Fall 2013 i Contents 1 Polytopes 1 1.1 Convex Combinations and V-Polytopes.....................
More informationLecture 3 Sections 2.2, 4.4. Mon, Aug 31, 2009
Model s Lecture 3 Sections 2.2, 4.4 World s Eye s Clip s s s Window s Hampden-Sydney College Mon, Aug 31, 2009 Outline Model s World s Eye s Clip s s s Window s 1 2 3 Model s World s Eye s Clip s s s Window
More informationIntroduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras
Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Module - 05 Lecture - 24 Solving LPs with mixed type of constraints In the
More informationLECTURES 3 and 4: Flows and Matchings
LECTURES 3 and 4: Flows and Matchings 1 Max Flow MAX FLOW (SP). Instance: Directed graph N = (V,A), two nodes s,t V, and capacities on the arcs c : A R +. A flow is a set of numbers on the arcs such that
More informationDiscrete Optimization. Lecture Notes 2
Discrete Optimization. Lecture Notes 2 Disjunctive Constraints Defining variables and formulating linear constraints can be straightforward or more sophisticated, depending on the problem structure. The
More informationChapter 3 Numerical Methods
Chapter 3 Numerical Methods Part 1 3.1 Linearization and Optimization of Functions of Vectors 1 Problem Notation 2 Outline 3.1.1 Linearization 3.1.2 Optimization of Objective Functions 3.1.3 Constrained
More informationAn introduction to pplex and the Simplex Method
An introduction to pplex and the Simplex Method Joanna Bauer Marc Bezem Andreas Halle November 16, 2012 Abstract Linear programs occur frequently in various important disciplines, such as economics, management,
More informationLecture 2 September 3
EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give
More information