3 Interior Point Method

Size: px
Start display at page:

Download "3 Interior Point Method"

Transcription

1 3 Interior Point Method Linear programming (LP) is one of the most useful mathematical techniques. Recent advances in computer technology and algorithms have improved computational speed by several orders of magnitude. Therefore, many problems that could not be addressed previously can now be successfully solved. One of the most well-nown LP methods is Simplex Method (SM), developed by Dantzig 34]. This method has remained competitive due to continuous improvements and research, and is considered very robust, general purpose, and efficient. It has been used to solve problems in operational research, business, and engineering. However, its polynomial complexity, with respect to the size of the problems, has led the scientific community to search for new algorithms with better guaranteed performance. In 1979, Khachiyan 35] developed the ellipsoid method and proved that LP problems belong to the class of polynomially solvable problems. His method reaches a solution in a time that is at worst polynomial in the problem size. However, although ideal for theoretical grounds, the algorithm failed to compete with the Simplex method. In 1984 Karmarar 36] proposed the projective algorithm which has better complexity bounds than Khachiyan s. Karmarar claimed that his method presents excellent performance for large problems when compared to SM. He did not show any results, but his wor originated a great deal of research activity. From , more than 1300 papers were published 37] in respect of this Interior Point Method (IPM) for LP. This method presents not only theoretical good performance but it is also very well suited for practical applications. SM and IPM wor in different ways. SM visits the vertices of the feasible region until it reaches the optimal, whereas IPM starts from the interior (or the exterior) of the feasible region and reaches the optimal through successive steps. Each SM iteration is computationally inexpensive, but the method usually requires many iterations. On the other hand, each IPM iteration is computationally expensive, but requires only a few iterations to converge.

2 Chapter 3. Interior Point Method 29 IPM is available in many commercial codes, such as CPLEX, LIPSOL or MATLAB. One of the most successful IPMs is the Primal-Dual IPM which is briefly discussed in the next section. 3.1 Optimality Conditions The LP problem in its standard form, also called the primal problem (PP) 38], can be expressed as: min x s.t. c T x Ax = b x 0, (3-1) where c, x R n ; b R m ; and A is an m n matrix with full row ran. Associated with the PP is the dual problem (DP): max y s.t. b T y A T y + z = c z 0, (3-2) where y R m, and z R n. The duality measure: µ = xt z n = ct x b T z n (3-3) is nonnegative and is zero only at optimal values (x, y ). It is commonly used as the test parameter to stop iteration. The Karush-Kuhn-Tucer conditions are: A T y + z Ax = c = b XZe = 0 (3-4) (x, z) 0, where X is a diagonal matrix with diagonal x, similarly for Z, and e is a vector of all ones with suitable dimension depending on the context. The vector (x, y, z ) is called a primal dual solution Primal-Dual Path Following The primal-dual methods find solutions (x, y, z ) by applying variants of Newton s method to the first three equations of Equation (3-4), so that x, z 0 are satisfied at every iteration. We can restate the optimality

3 Chapter 3. Interior Point Method 30 conditions in terms of the mapping F : R 2n+m R 2n+m : A T y + z c F (x, y, z) = Ax b = 0 XZe (3-5) (x, z) 0. Newton s method is used to obtain the next iteration. It consists of obtaining the search direction ( x, y, z) by solving: x J (x, y, z) y = F (x, y, z), (3-6) z where J is the Jacobian of F. If r b = Ax b and r c = A T y + z c, then: 0 A T I x r c A 0 0 y = r b. (3-7) Z 0 X z XZe The next feasible iteration is: (x, y, z) + α ( x, y, z), (3-8) where parameter α (0, 1] is chosen such that the next iteration is ept strictly feasible, i.e., (x, z) > 0. Usually, we tae a step toward a point such that x i s i = σµ, where µ is the current duality measure and α, σ 0, 1] a. Figure 3.1 shows the steps from an initial point to the optimized one. The modified step equation is then: 0 A T I x A 0 0 y = Z 0 X z r c r b XZe + σµe. (3-9) The Central Path The central path is the arc of points C parameterized by a scalar τ > 0. Each point of (x τ, y τ, z τ ) C satisfies: a Further details on σ will be provided later in this chapter.

4 Chapter 3. Interior Point Method 31 Figure 3.1: Steps towards the optimum in the x space. A T y + z = c Ax = b XZe = τe (x, z) 0. (3-10) Equations (3-10) correspond to the optimality conditions for a logarithmic barrier version of the primal problem. This is obtained by adding a barrier term which penalizes non-positive values. The barrier problem is defined as: min c T x τ n i=1 ln x i s.t. Ax = b. (3-11) Note that equations (3-11) approximate the original system (3-4) as τ 0. Then the central path gives a route to reach the optimum. Figure 3.2 shows contour lines for the logarithmic barrier problem for different values of τ, where the dots indicate optimum values of τ. The σ parameter is called the centering parameter and values close to 1 give directions closer to the central path C. A zero value gives the standard Newton step. Practical algorithms tend to search an equilibrium between centrality and taing larger steps to the optimum by carefully choosing σ and α Long-Step Path Following In this algorithm the α parameter in each iteration is chosen to be the largest value to satisfy x + α x 0, for primal variables (α pri ), and z + α z 0, for dual variables (α dual ).

5 Chapter 3. Interior Point Method 32 Figure 3.2: Central path and contours. (a) τ = 100. (b) τ = 3. (c) τ = 1. (d) τ = Matlab code for maing this graphic is showed in Appendix A. The η (η (0, 1)) parameter is usually set very close to 1 to prevent zero values for x and z, while at the same time maximizing the step lengths Predictor-Corrector Algorithm As discussed above, the centering parameter, σ, has the effect of producing iterations closer to the central path (if σ 0) with slow convergence or far from the central path (if σ 1) with faster convergence, but close to the boundaries of the feasible region. The predictor-corrector algorithm attempts to balance these behaviors to accelerate the convergence. The predictor-corrector finds σ in an adaptive way. First, the Newton s search direction is found by solving Equation (3-15): the predictor step. The maximum α values are found from Equation (3-16) and σ is calculated according to the heuristic σ = (µ aff /µ) 3. The centering parameter, σ, is included in Equation (3-17) for the corrector step, and the search direction is also obtained. The predictor-corrector needs to solve two linear systems

6 Chapter 3. Interior Point Method 33 Algorithm 1 Long-Step IPM 1: Initialization 2: Given σ 0, 1] and (x 0, y 0, z 0 ) 3: for = 0, 1, 2,... do 4: Set µ = (x ) T z /n 5: Solve to obtain ( x, y, z ) 6: Choose α pri 7: Set 8: end for 0 AT I A 0 0 Z 0 X α pri,max = α pri and α dual min x y = z such that x i i: x i <0 x i = min(1, η α pri,max ),, α dual,max = αdual r c r b XZe + σµe min z i i: z i <0 z i (3-12) = min(1, η α dual,max ) (3-13) ( x +1 = x + α pri x y +1, z +1) = ( y, z ) ( + α dual y, z ) (3-14) with the same coefficient matrix but different right hand terms. Direct solvers deal with this by factorizing once and maing two bac substitutions at low cost Solving the linear systems As discussed above, Equations (3-12), (3-15) and (3-17) only differ on the right hand side and they may be expressed more generally as: 0 A T I x r c A 0 0 y = r b. (3-20) Z 0 X z r xz Since x and z are positive, X and Z are invertible, and Equations (3-20) can be rewritten as: D 2 xz A 0 A T ] ] x y ] r c + X 1 r xz = r b z = X 1 r xz X 1 Z x, (3-21) where D xz = Z 1/2 X 1/2. This is nown as the augmented system and can be expressed as:

7 Chapter 3. Interior Point Method 34 Algorithm 2 Predictor-Corrector IPM 1: Initialization 2: Given (x 0, y 0, z 0 ) 3: for = 0, 1, 2,... do 4: Solve to obtain ( x aff, y aff, z aff) 5: Choose α pri aff, αdual aff α pri aff α dual aff 0 AT I A 0 0 Z 0 X x aff y aff = z aff and µ aff such that ( = min 1, min ( = min i: x aff i <0 1, min i: z aff i <0 x i x aff i z i z aff i µ aff = (x + α pri aff xaff ) T (z + α dual ) ) r c r b (3-15) XZe aff zaff )/n (3-16) 6: Set σ = (µ aff /µ) 3 7: Solve to obtain ( x, y, z) 0 AT I x r c A 0 0 y = r b (3-17) Z 0 X z XZe X aff Z aff + σµe 8: Calculate α pri, αdual 9: Set 10: end for α pri = min(1, η α pri,max ), αdual = min(1, η α dual,max ) (3-18) ( x +1 = x + α pri x y +1, z +1) = ( y, z ) ( + α dual y, z ) (3-19) AD 2 xza T y = r b AXZ 1 r c + AZ 1 r xz z = r c A T y (3-22) x = Z 1 r xz XZ 1 z. Equations 3-22 are the normal equations. The three systems of equations are equivalent, but have different characteristics. Usually, system (3-20) is not solved. The augmented system (Equations 3-21) is more suitable to solve. The normal equations (Equations 3-22) are attractive because of their compact form and also because the matrix AD 2 xza T is positive definite, allowing the use of nown direct and iterative solvers. If a direct solver is used, the term AD 2 xza T can be factorized by the Cholesy algorithm and then a bac substitution is applied to obtain y. However, D xz contains many zeros and also large numbers, which can lead to AD 2 xza T being highly ill-conditioned or singular

8 Chapter 3. Interior Point Method 35 at the final iterations. The Cholesy algorithm needs to be modified to address this problem. If the truss optimization problem of the previous chapter (see Equation 2-5) is solved using IPM, the following assignments can be made: ] A = B B b = f c = l σ Tl ] (3-23) where the term B B x = σ C s + s ], ] is replaced by AD 2 xza T, and the diagonal matrix D 2 xz is partitioned into four submatrices of equal size: ] ] ] AD 2 xza T D xz,1 0 B T = B B 0 D xz,2 B T. The product term in the right hand side can be reduced to: (3-24) M = BD ipm B T, (3-25) where D ipm = D 2 xz,1 + D 2 xz,2. This is the form that will be used in the solvers described in the next chapters. Note that B is sparse and constant, but D ipm varies in every IPM iteration. As the iterations proceed, this term has both small and large numbers and BD ipm B T may become ill-conditioned. This causes numerical problems for direct and iterative solvers, although iterative solvers are more affected. When iterative solvers, such as PCG, are used it is not necessary to explicitly assemble BD ipm B T because these methods rely on matrix vector products of the form BD ipm B T v, where v is a given vector. An important aspect of the implementation presented here is assembling B and BD ipm B T, which is the subject of the next chapter.

ORIE 6300 Mathematical Programming I November 13, Lecture 23. max b T y. x 0 s 0. s.t. A T y + s = c

ORIE 6300 Mathematical Programming I November 13, Lecture 23. max b T y. x 0 s 0. s.t. A T y + s = c ORIE 63 Mathematical Programming I November 13, 214 Lecturer: David P. Williamson Lecture 23 Scribe: Mukadder Sevi Baltaoglu 1 Interior Point Methods Consider the standard primal and dual linear programs:

More information

Programming, numerics and optimization

Programming, numerics and optimization Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June

More information

A PRIMAL-DUAL EXTERIOR POINT ALGORITHM FOR LINEAR PROGRAMMING PROBLEMS

A PRIMAL-DUAL EXTERIOR POINT ALGORITHM FOR LINEAR PROGRAMMING PROBLEMS Yugoslav Journal of Operations Research Vol 19 (2009), Number 1, 123-132 DOI:10.2298/YUJOR0901123S A PRIMAL-DUAL EXTERIOR POINT ALGORITHM FOR LINEAR PROGRAMMING PROBLEMS Nikolaos SAMARAS Angelo SIFELARAS

More information

EARLY INTERIOR-POINT METHODS

EARLY INTERIOR-POINT METHODS C H A P T E R 3 EARLY INTERIOR-POINT METHODS An interior-point algorithm is one that improves a feasible interior solution point of the linear program by steps through the interior, rather than one that

More information

Submodularity Reading Group. Matroid Polytopes, Polymatroid. M. Pawan Kumar

Submodularity Reading Group. Matroid Polytopes, Polymatroid. M. Pawan Kumar Submodularity Reading Group Matroid Polytopes, Polymatroid M. Pawan Kumar http://www.robots.ox.ac.uk/~oval/ Outline Linear Programming Matroid Polytopes Polymatroid Polyhedron Ax b A : m x n matrix b:

More information

Programs. Introduction

Programs. Introduction 16 Interior Point I: Linear Programs Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen

More information

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Mathematical and Algorithmic Foundations Linear Programming and Matchings Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

16.410/413 Principles of Autonomy and Decision Making

16.410/413 Principles of Autonomy and Decision Making 16.410/413 Principles of Autonomy and Decision Making Lecture 17: The Simplex Method Emilio Frazzoli Aeronautics and Astronautics Massachusetts Institute of Technology November 10, 2010 Frazzoli (MIT)

More information

Matrix-free IPM with GPU acceleration

Matrix-free IPM with GPU acceleration Matrix-free IPM with GPU acceleration Julian Hall, Edmund Smith and Jacek Gondzio School of Mathematics University of Edinburgh jajhall@ed.ac.uk 29th June 2011 Linear programming theory Primal-dual pair

More information

AN EXPERIMENTAL INVESTIGATION OF A PRIMAL- DUAL EXTERIOR POINT SIMPLEX ALGORITHM

AN EXPERIMENTAL INVESTIGATION OF A PRIMAL- DUAL EXTERIOR POINT SIMPLEX ALGORITHM AN EXPERIMENTAL INVESTIGATION OF A PRIMAL- DUAL EXTERIOR POINT SIMPLEX ALGORITHM Glavelis Themistoklis Samaras Nikolaos Paparrizos Konstantinos PhD Candidate Assistant Professor Professor Department of

More information

Simulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016

Simulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016 Simulation Lecture O Optimization: Linear Programming Saeed Bastani April 06 Outline of the course Linear Programming ( lecture) Integer Programming ( lecture) Heuristics and Metaheursitics (3 lectures)

More information

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case

Ellipsoid Algorithm :Algorithms in the Real World. Ellipsoid Algorithm. Reduction from general case Ellipsoid Algorithm 15-853:Algorithms in the Real World Linear and Integer Programming II Ellipsoid algorithm Interior point methods First polynomial-time algorithm for linear programming (Khachian 79)

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

LECTURE 6: INTERIOR POINT METHOD. 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm

LECTURE 6: INTERIOR POINT METHOD. 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm LECTURE 6: INTERIOR POINT METHOD 1. Motivation 2. Basic concepts 3. Primal affine scaling algorithm 4. Dual affine scaling algorithm Motivation Simplex method works well in general, but suffers from exponential-time

More information

Interior Point I. Lab 21. Introduction

Interior Point I. Lab 21. Introduction Lab 21 Interior Point I Lab Objective: For decades after its invention, the Simplex algorithm was the only competitive method for linear programming. The past 30 years, however, have seen the discovery

More information

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014 5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38

More information

Chapter 15 Introduction to Linear Programming

Chapter 15 Introduction to Linear Programming Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of

More information

Figure 6.1: Truss topology optimization diagram.

Figure 6.1: Truss topology optimization diagram. 6 Implementation 6.1 Outline This chapter shows the implementation details to optimize the truss, obtained in the ground structure approach, according to the formulation presented in previous chapters.

More information

Convex Optimization CMU-10725

Convex Optimization CMU-10725 Convex Optimization CMU-10725 Ellipsoid Methods Barnabás Póczos & Ryan Tibshirani Outline Linear programs Simplex algorithm Running time: Polynomial or Exponential? Cutting planes & Ellipsoid methods for

More information

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization

Performance Evaluation of an Interior Point Filter Line Search Method for Constrained Optimization 6th WSEAS International Conference on SYSTEM SCIENCE and SIMULATION in ENGINEERING, Venice, Italy, November 21-23, 2007 18 Performance Evaluation of an Interior Point Filter Line Search Method for Constrained

More information

Optimized Choice of Parameters in interior-point methods for linear programming. Luiz-Rafael dos Santos

Optimized Choice of Parameters in interior-point methods for linear programming. Luiz-Rafael dos Santos Optimized Choice of Parameters in interior-point methods for linear programming Luiz-Rafael dos Santos Joint work with F. Villas-Bôas, A. Oliveira and C. Perin 1st Brazilian Workshop on Interior Point

More information

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Lecture 18 All-Integer Dual Algorithm We continue the discussion on the all integer

More information

Theoretical Concepts of Machine Learning

Theoretical Concepts of Machine Learning Theoretical Concepts of Machine Learning Part 2 Institute of Bioinformatics Johannes Kepler University, Linz, Austria Outline 1 Introduction 2 Generalization Error 3 Maximum Likelihood 4 Noise Models 5

More information

Code Generation for Embedded Convex Optimization

Code Generation for Embedded Convex Optimization Code Generation for Embedded Convex Optimization Jacob Mattingley Stanford University October 2010 Convex optimization Problems solvable reliably and efficiently Widely used in scheduling, finance, engineering

More information

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles

Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles INTERNATIONAL JOURNAL OF MATHEMATICS MODELS AND METHODS IN APPLIED SCIENCES Comparison of Interior Point Filter Line Search Strategies for Constrained Optimization by Performance Profiles M. Fernanda P.

More information

Solving the Master Linear Program in Column Generation Algorithms for Airline Crew Scheduling using a Subgradient Method.

Solving the Master Linear Program in Column Generation Algorithms for Airline Crew Scheduling using a Subgradient Method. Solving the Master Linear Program in Column Generation Algorithms for Airline Crew Scheduling using a Subgradient Method Per Sjögren November 28, 2009 Abstract A subgradient method for solving large linear

More information

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize. Cornell University, Fall 2017 CS 6820: Algorithms Lecture notes on the simplex method September 2017 1 The Simplex Method We will present an algorithm to solve linear programs of the form maximize subject

More information

Crash-Starting the Simplex Method

Crash-Starting the Simplex Method Crash-Starting the Simplex Method Ivet Galabova Julian Hall School of Mathematics, University of Edinburgh Optimization Methods and Software December 2017 Ivet Galabova, Julian Hall Crash-Starting Simplex

More information

(67686) Mathematical Foundations of AI July 30, Lecture 11

(67686) Mathematical Foundations of AI July 30, Lecture 11 (67686) Mathematical Foundations of AI July 30, 2008 Lecturer: Ariel D. Procaccia Lecture 11 Scribe: Michael Zuckerman and Na ama Zohary 1 Cooperative Games N = {1,...,n} is the set of players (agents).

More information

Mathematical Programming and Research Methods (Part II)

Mathematical Programming and Research Methods (Part II) Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types

More information

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg MVE165/MMG630, Integer linear programming algorithms Ann-Brith Strömberg 2009 04 15 Methods for ILP: Overview (Ch. 14.1) Enumeration Implicit enumeration: Branch and bound Relaxations Decomposition methods:

More information

Gate Sizing by Lagrangian Relaxation Revisited

Gate Sizing by Lagrangian Relaxation Revisited Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007

More information

February 19, Integer programming. Outline. Problem formulation. Branch-andbound

February 19, Integer programming. Outline. Problem formulation. Branch-andbound Olga Galinina olga.galinina@tut.fi ELT-53656 Network Analysis and Dimensioning II Department of Electronics and Communications Engineering Tampere University of Technology, Tampere, Finland February 19,

More information

High performance computing and the simplex method

High performance computing and the simplex method Julian Hall, Qi Huangfu and Edmund Smith School of Mathematics University of Edinburgh 12th April 2011 The simplex method for LP Not... Nonlinear programming... Integer programming... Stochastic programming......

More information

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.

More information

ME 391Q Network Flow Programming

ME 391Q Network Flow Programming ME 9Q Network Flow Programming Final Exam, Summer 00. ( Points) The figure below shows an undirected network. The parameters on the edges are the edge lengths. Find the shortest path tree using Dijkstra

More information

Integer Programming Theory

Integer Programming Theory Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x

More information

DEGENERACY AND THE FUNDAMENTAL THEOREM

DEGENERACY AND THE FUNDAMENTAL THEOREM DEGENERACY AND THE FUNDAMENTAL THEOREM The Standard Simplex Method in Matrix Notation: we start with the standard form of the linear program in matrix notation: (SLP) m n we assume (SLP) is feasible, and

More information

NATCOR Convex Optimization Linear Programming 1

NATCOR Convex Optimization Linear Programming 1 NATCOR Convex Optimization Linear Programming 1 Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk 5 June 2018 What is linear programming (LP)? The most important model used in

More information

3 INTEGER LINEAR PROGRAMMING

3 INTEGER LINEAR PROGRAMMING 3 INTEGER LINEAR PROGRAMMING PROBLEM DEFINITION Integer linear programming problem (ILP) of the decision variables x 1,..,x n : (ILP) subject to minimize c x j j n j= 1 a ij x j x j 0 x j integer n j=

More information

Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity

Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity Marc Uetz University of Twente m.uetz@utwente.nl Lecture 5: sheet 1 / 26 Marc Uetz Discrete Optimization Outline 1 Min-Cost Flows

More information

Convex Optimization M2

Convex Optimization M2 Convex Optimization M2 Lecture 1 A. d Aspremont. Convex Optimization M2. 1/49 Today Convex optimization: introduction Course organization and other gory details... Convex sets, basic definitions. A. d

More information

Algorithms for convex optimization

Algorithms for convex optimization Algorithms for convex optimization Michal Kočvara Institute of Information Theory and Automation Academy of Sciences of the Czech Republic and Czech Technical University kocvara@utia.cas.cz http://www.utia.cas.cz/kocvara

More information

Notes for Lecture 18

Notes for Lecture 18 U.C. Berkeley CS17: Intro to CS Theory Handout N18 Professor Luca Trevisan November 6, 21 Notes for Lecture 18 1 Algorithms for Linear Programming Linear programming was first solved by the simplex method

More information

Math 5593 Linear Programming Lecture Notes

Math 5593 Linear Programming Lecture Notes Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................

More information

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited.

Contents. I Basics 1. Copyright by SIAM. Unauthorized reproduction of this article is prohibited. page v Preface xiii I Basics 1 1 Optimization Models 3 1.1 Introduction... 3 1.2 Optimization: An Informal Introduction... 4 1.3 Linear Equations... 7 1.4 Linear Optimization... 10 Exercises... 12 1.5

More information

Sparse Convex Optimization on GPUs

Sparse Convex Optimization on GPUs Sparse Convex Optimization on GPUs by Marco Maggioni B.A. (Politecnico di Milano) 2006 M.S. (University of Illinois at Chicago) 2008 M.S. (Politecnico di Milano) 2010 M.S. (University of Illinois at Chicago)

More information

What is linear programming (LP)? NATCOR Convex Optimization Linear Programming 1. Solving LP problems: The standard simplex method

What is linear programming (LP)? NATCOR Convex Optimization Linear Programming 1. Solving LP problems: The standard simplex method NATCOR Convex Optimization Linear Programming 1 Julian Hall School of Mathematics University of Edinburgh jajhall@ed.ac.uk 14 June 2016 What is linear programming (LP)? The most important model used in

More information

Introduction. Linear because it requires linear functions. Programming as synonymous of planning.

Introduction. Linear because it requires linear functions. Programming as synonymous of planning. LINEAR PROGRAMMING Introduction Development of linear programming was among the most important scientific advances of mid-20th cent. Most common type of applications: allocate limited resources to competing

More information

A specialized interior-point algorithm for huge minimum convex cost flows in bipartite networks

A specialized interior-point algorithm for huge minimum convex cost flows in bipartite networks A specialized interior-point algorithm for huge minimum convex cost flows in bipartite networks Jordi Castro Dept. of Stat. and Oper. Res. Universitat Politècnica de Catalunya Barcelona, Catalonia jordi.castro@upc.edu

More information

COLUMN GENERATION IN LINEAR PROGRAMMING

COLUMN GENERATION IN LINEAR PROGRAMMING COLUMN GENERATION IN LINEAR PROGRAMMING EXAMPLE: THE CUTTING STOCK PROBLEM A certain material (e.g. lumber) is stocked in lengths of 9, 4, and 6 feet, with respective costs of $5, $9, and $. An order for

More information

Linear Programming Duality and Algorithms

Linear Programming Duality and Algorithms COMPSCI 330: Design and Analysis of Algorithms 4/5/2016 and 4/7/2016 Linear Programming Duality and Algorithms Lecturer: Debmalya Panigrahi Scribe: Tianqi Song 1 Overview In this lecture, we will cover

More information

Some Advanced Topics in Linear Programming

Some Advanced Topics in Linear Programming Some Advanced Topics in Linear Programming Matthew J. Saltzman July 2, 995 Connections with Algebra and Geometry In this section, we will explore how some of the ideas in linear programming, duality theory,

More information

COT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748

COT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748 COT 6936: Topics in Algorithms! Giri Narasimhan ECS 254A / EC 2443; Phone: x3748 giri@cs.fiu.edu http://www.cs.fiu.edu/~giri/teach/cot6936_s12.html https://moodle.cis.fiu.edu/v2.1/course/view.php?id=174

More information

Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem

Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem CENTER FOR MACHINE PERCEPTION CZECH TECHNICAL UNIVERSITY Sequential Coordinate-wise Algorithm for Non-negative Least Squares Problem Woring document of the EU project COSPAL IST-004176 Vojtěch Franc, Miro

More information

Linear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage).

Linear Programming. Readings: Read text section 11.6, and sections 1 and 2 of Tom Ferguson s notes (see course homepage). Linear Programming Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory: Feasible Set, Vertices, Existence of Solutions. Equivalent formulations. Outline

More information

Solving Linear Programs Using the Simplex Method (Manual)

Solving Linear Programs Using the Simplex Method (Manual) Solving Linear Programs Using the Simplex Method (Manual) GáborRétvári E-mail: retvari@tmit.bme.hu The GNU Octave Simplex Solver Implementation As part of the course material two simple GNU Octave/MATLAB

More information

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued.

Linear Programming. Widget Factory Example. Linear Programming: Standard Form. Widget Factory Example: Continued. Linear Programming Widget Factory Example Learning Goals. Introduce Linear Programming Problems. Widget Example, Graphical Solution. Basic Theory:, Vertices, Existence of Solutions. Equivalent formulations.

More information

Solution Methods Numerical Algorithms

Solution Methods Numerical Algorithms Solution Methods Numerical Algorithms Evelien van der Hurk DTU Managment Engineering Class Exercises From Last Time 2 DTU Management Engineering 42111: Static and Dynamic Optimization (6) 09/10/2017 Class

More information

Research Interests Optimization:

Research Interests Optimization: Mitchell: Research interests 1 Research Interests Optimization: looking for the best solution from among a number of candidates. Prototypical optimization problem: min f(x) subject to g(x) 0 x X IR n Here,

More information

Linear Programming. Linear programming provides methods for allocating limited resources among competing activities in an optimal way.

Linear Programming. Linear programming provides methods for allocating limited resources among competing activities in an optimal way. University of Southern California Viterbi School of Engineering Daniel J. Epstein Department of Industrial and Systems Engineering ISE 330: Introduction to Operations Research - Deterministic Models Fall

More information

New Directions in Linear Programming

New Directions in Linear Programming New Directions in Linear Programming Robert Vanderbei November 5, 2001 INFORMS Miami Beach NOTE: This is a talk mostly on pedagogy. There will be some new results. It is not a talk on state-of-the-art

More information

Three nearly scaling-invariant versions of an exterior point algorithm for linear programming

Three nearly scaling-invariant versions of an exterior point algorithm for linear programming Optimization A Journal of Mathematical Programming and Operations Research ISSN: 0233-1934 (Print) 1029-4945 (Online) Journal homepage: http://www.tandfonline.com/loi/gopt20 Three nearly scaling-invariant

More information

Computational issues in linear programming

Computational issues in linear programming Computational issues in linear programming Julian Hall School of Mathematics University of Edinburgh 15th May 2007 Computational issues in linear programming Overview Introduction to linear programming

More information

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming Linear Programming 3 describes a broad class of optimization tasks in which both the optimization criterion and the constraints are linear functions. Linear Programming consists of three parts: A set of

More information

Detecting Infeasibility in Infeasible-Interior-Point. Methods for Optimization

Detecting Infeasibility in Infeasible-Interior-Point. Methods for Optimization FOCM 02 Infeasible Interior Point Methods 1 Detecting Infeasibility in Infeasible-Interior-Point Methods for Optimization Slide 1 Michael J. Todd, School of Operations Research and Industrial Engineering,

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Constrained Optimization Marc Toussaint U Stuttgart Constrained Optimization General constrained optimization problem: Let R n, f : R n R, g : R n R m, h : R n R l find min

More information

LOAD SHEDDING AN EFFICIENT USE OF LTC TRANSFORMERS

LOAD SHEDDING AN EFFICIENT USE OF LTC TRANSFORMERS LOAD SHEDDING AN EFFICIENT USE OF LTC TRANSFORMERS Luciano V. Barboza André A. P. Lerm Roberto S. Salgado Catholic University of Pelotas Federal Center for Technological Education of Pelotas Federal University

More information

1. Lecture notes on bipartite matching February 4th,

1. Lecture notes on bipartite matching February 4th, 1. Lecture notes on bipartite matching February 4th, 2015 6 1.1.1 Hall s Theorem Hall s theorem gives a necessary and sufficient condition for a bipartite graph to have a matching which saturates (or matches)

More information

Section Notes 4. Duality, Sensitivity, and the Dual Simplex Algorithm. Applied Math / Engineering Sciences 121. Week of October 8, 2018

Section Notes 4. Duality, Sensitivity, and the Dual Simplex Algorithm. Applied Math / Engineering Sciences 121. Week of October 8, 2018 Section Notes 4 Duality, Sensitivity, and the Dual Simplex Algorithm Applied Math / Engineering Sciences 121 Week of October 8, 2018 Goals for the week understand the relationship between primal and dual

More information

Network optimization: An overview

Network optimization: An overview Network optimization: An overview Mathias Johanson Alkit Communications 1 Introduction Various kinds of network optimization problems appear in many fields of work, including telecommunication systems,

More information

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization? Linear and Integer Programming 15-853:Algorithms in the Real World Linear and Integer Programming I Introduction Geometric Interpretation Simplex Method Linear or Integer programming maximize z = c T x

More information

An example of LP problem: Political Elections

An example of LP problem: Political Elections Linear Programming An example of LP problem: Political Elections Suppose that you are a politician trying to win an election. Your district has three different types of areas: urban, suburban, and rural.

More information

Support Vector Machines

Support Vector Machines Support Vector Machines SVM Discussion Overview. Importance of SVMs. Overview of Mathematical Techniques Employed 3. Margin Geometry 4. SVM Training Methodology 5. Overlapping Distributions 6. Dealing

More information

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS GRADO EN A.D.E. GRADO EN ECONOMÍA GRADO EN F.Y.C. ACADEMIC YEAR 2011-12 INDEX UNIT 1.- AN INTRODUCCTION TO OPTIMIZATION 2 UNIT 2.- NONLINEAR PROGRAMMING

More information

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points. 5.93 Optimization Methods Final Examination Instructions:. There are 5 problems each w i t h 2 p o i n ts for a maximum of points. 2. You are allowed to use class notes, your homeworks, solutions to homework

More information

A New Unblocking Technique to Warmstart Interior Point Methods based on Sensitivity Analysis. December 19th, 2006 revised June 1, 2007 MS

A New Unblocking Technique to Warmstart Interior Point Methods based on Sensitivity Analysis. December 19th, 2006 revised June 1, 2007 MS A New Unblocking Technique to Warmstart Interior Point Methods based on Sensitivity Analysis Jacek Gondzio Andreas Grothey December 19th, 2006 revised June 1, 2007 MS-06-005 1 A New Unblocking Technique

More information

arxiv: v1 [cs.cc] 30 Jun 2017

arxiv: v1 [cs.cc] 30 Jun 2017 On the Complexity of Polytopes in LI( Komei Fuuda May Szedlá July, 018 arxiv:170610114v1 [cscc] 30 Jun 017 Abstract In this paper we consider polytopes given by systems of n inequalities in d variables,

More information

The Affine Scaling Method

The Affine Scaling Method MA33 Linear Programming W. J. Martin October 9, 8 The Affine Scaling Method Overview Given a linear programming problem in equality form with full rank constraint matrix and a strictly positive feasible

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,

More information

Generalized Network Flow Programming

Generalized Network Flow Programming Appendix C Page Generalized Network Flow Programming This chapter adapts the bounded variable primal simplex method to the generalized minimum cost flow problem. Generalized networks are far more useful

More information

A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems

A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems E. O. Effanga Department of Mathematics/Statistics and Comp. Science University of Calabar P.M.B. 1115, Calabar, Cross

More information

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Lecture 16 Cutting Plane Algorithm We shall continue the discussion on integer programming,

More information

1. Lecture notes on bipartite matching

1. Lecture notes on bipartite matching Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans February 5, 2017 1. Lecture notes on bipartite matching Matching problems are among the fundamental problems in

More information

(Sparse) Linear Solvers

(Sparse) Linear Solvers (Sparse) Linear Solvers Ax = B Why? Many geometry processing applications boil down to: solve one or more linear systems Parameterization Editing Reconstruction Fairing Morphing 2 Don t you just invert

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

Characterizing Improving Directions Unconstrained Optimization

Characterizing Improving Directions Unconstrained Optimization Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not

More information

OPERATIONS RESEARCH. Linear Programming Problem

OPERATIONS RESEARCH. Linear Programming Problem OPERATIONS RESEARCH Chapter 1 Linear Programming Problem Prof. Bibhas C. Giri Department of Mathematics Jadavpur University Kolkata, India Email: bcgiri.jumath@gmail.com 1.0 Introduction Linear programming

More information

MPC Toolbox with GPU Accelerated Optimization Algorithms

MPC Toolbox with GPU Accelerated Optimization Algorithms Downloaded from orbit.dtu.dk on: Apr 21, 218 MPC Toolbox with GPU Accelerated Optimization Algorithms Gade-Nielsen, Nicolai Fog; Jørgensen, John Bagterp; Dammann, Bernd Published in: The 1th European Workshop

More information

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i

More information

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D.

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D. Linear Programming Module Outline Introduction The Linear Programming Model Examples of Linear Programming Problems Developing Linear Programming Models Graphical Solution to LP Problems The Simplex Method

More information

Linear Optimization and Extensions: Theory and Algorithms

Linear Optimization and Extensions: Theory and Algorithms AT&T Linear Optimization and Extensions: Theory and Algorithms Shu-Cherng Fang North Carolina State University Sarai Puthenpura AT&T Bell Labs Prentice Hall, Englewood Cliffs, New Jersey 07632 Contents

More information

Linear Programming Problems

Linear Programming Problems Linear Programming Problems Two common formulations of linear programming (LP) problems are: min Subject to: 1,,, 1,2,,;, max Subject to: 1,,, 1,2,,;, Linear Programming Problems The standard LP problem

More information

Linear Programming 1

Linear Programming 1 Linear Programming 1 Fei Li March 5, 2012 1 With references of Algorithms by S. Dasgupta, C. H. Papadimitriou, and U. V. Vazirani. Many of the problems for which we want algorithms are optimization tasks.

More information

Size of a problem instance: Bigger instances take

Size of a problem instance: Bigger instances take 2.1 Integer Programming and Combinatorial Optimization Slide set 2: Computational Complexity Katta G. Murty Lecture slides Aim: To study efficiency of various algo. for solving problems, and to classify

More information

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm Instructor: Shaddin Dughmi Algorithms for Convex Optimization We will look at 2 algorithms in detail: Simplex and Ellipsoid.

More information

WE consider the gate-sizing problem, that is, the problem

WE consider the gate-sizing problem, that is, the problem 2760 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I: REGULAR PAPERS, VOL 55, NO 9, OCTOBER 2008 An Efficient Method for Large-Scale Gate Sizing Siddharth Joshi and Stephen Boyd, Fellow, IEEE Abstract We consider

More information

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Introduction to Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Module 03 Simplex Algorithm Lecture - 03 Tabular form (Minimization) In this

More information