On Clarkson s Las Vegas Algorithms for Linear and Integer Programming When the Dimension is Small

Similar documents
Integer Programming Theory

Trapezoidal decomposition:

A Subexponential Randomized Simplex Algorithm

Linear Programming in Small Dimensions

5.3 Cutting plane methods and Gomory fractional cuts

Convex Geometry arising in Optimization

Treewidth and graph minors

/ Approximation Algorithms Lecturer: Michael Dinitz Topic: Linear Programming Date: 2/24/15 Scribe: Runze Tang

Computational Geometry: Lecture 5

9 Bounds for the Knapsack Problem (March 6)

The Simplex Algorithm

Lecture 2 - Introduction to Polytopes

Convex Optimization CMU-10725

Integral Geometry and the Polynomial Hirsch Conjecture

Polytopes Course Notes

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

CS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm. Instructor: Shaddin Dughmi

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

Optimality certificates for convex minimization and Helly numbers

Math 5593 Linear Programming Lecture Notes

Simplicial Complexes: Second Lecture

15-451/651: Design & Analysis of Algorithms November 4, 2015 Lecture #18 last changed: November 22, 2015

Optimality certificates for convex minimization and Helly numbers

LP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008

11 Linear Programming

CS 580: Algorithm Design and Analysis. Jeremiah Blocki Purdue University Spring 2018

Approximation Algorithms

6. Lecture notes on matroid intersection

THEORY OF LINEAR AND INTEGER PROGRAMMING

Optimization of Design. Lecturer:Dung-An Wang Lecture 8

Introduction to Mathematical Programming IE406. Lecture 20. Dr. Ted Ralphs

CS675: Convex and Combinatorial Optimization Spring 2018 Convex Sets. Instructor: Shaddin Dughmi

Linear programming and duality theory

However, this is not always true! For example, this fails if both A and B are closed and unbounded (find an example).

Lecture 3. Corner Polyhedron, Intersection Cuts, Maximal Lattice-Free Convex Sets. Tepper School of Business Carnegie Mellon University, Pittsburgh

Coloring 3-Colorable Graphs

COMP331/557. Chapter 2: The Geometry of Linear Programming. (Bertsimas & Tsitsiklis, Chapter 2)

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Dual-fitting analysis of Greedy for Set Cover

DM545 Linear and Integer Programming. Lecture 2. The Simplex Method. Marco Chiarandini

Greedy Algorithms 1 {K(S) K(S) C} For large values of d, brute force search is not feasible because there are 2 d {1,..., d}.

11. APPROXIMATION ALGORITHMS

Financial Optimization ISE 347/447. Lecture 13. Dr. Ted Ralphs

Maximal Independent Set

1 Linear Programming. 1.1 Optimizion problems and convex polytopes 1 LINEAR PROGRAMMING

Notes for Lecture 24

2. Optimization problems 6

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs

15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018

ACTUALLY DOING IT : an Introduction to Polyhedral Computation

CMPSCI611: The Simplex Algorithm Lecture 24

Lecture notes on the simplex method September We will present an algorithm to solve linear programs of the form. maximize.

Maximal Independent Set

Math 414 Lecture 2 Everyone have a laptop?

Combinatorial Geometry & Topology arising in Game Theory and Optimization

Lecture : Topological Space

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs

11. APPROXIMATION ALGORITHMS

Lecture 15: The subspace topology, Closed sets

3 INTEGER LINEAR PROGRAMMING

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg

Mathematical Programming and Research Methods (Part II)

Theorem 2.9: nearest addition algorithm

Framework for Design of Dynamic Programming Algorithms

An iteration of the simplex method (a pivot )

Lecture 12: Feasible direction methods

Stable sets, corner polyhedra and the Chvátal closure

Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity

Lecture 4: Linear Programming

! Linear programming"! Duality "! Smallest enclosing disk"

LECTURES 3 and 4: Flows and Matchings

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 4: Convex Sets. Instructor: Shaddin Dughmi

Chapter 15 Introduction to Linear Programming

Open and Closed Sets

Compact Sets. James K. Peterson. September 15, Department of Biological Sciences and Department of Mathematical Sciences Clemson University

POLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if

3 No-Wait Job Shops with Variable Processing Times

k-selection Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong

15. Cutting plane and ellipsoid methods

Distance Sequences in Locally Infinite Vertex-Transitive Digraphs

Some Advanced Topics in Linear Programming

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

4 Integer Linear Programming (ILP)

The complement of PATH is in NL

Greedy Algorithms and Matroids. Andreas Klappenecker

Greedy Algorithms 1. For large values of d, brute force search is not feasible because there are 2 d

From final point cuts to!-polyhedral cuts

Lecture 2: August 31

Technische Universität München Zentrum Mathematik

11. APPROXIMATION ALGORITHMS

Exact adaptive parallel algorithms for data depth problems. Vera Rosta Department of Mathematics and Statistics McGill University, Montreal

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

Week 5. Convex Optimization

Monotone Paths in Geometric Triangulations

Lecture The Ellipsoid Algorithm

A CSP Search Algorithm with Reduced Branching Factor

On Covering a Graph Optimally with Induced Subgraphs

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18

Lecture 5: Properties of convex sets

9.5 Equivalence Relations

Transcription:

On Clarkson s Las Vegas Algorithms for Linear and Integer Programming When the Dimension is Small Robert Bassett March 10, 2014

2 Question/Why Do We Care Motivating Question: Given a linear or integer program with the condition that the number of variables is small, can we come up with algorithm(s) that are particularly well-suited to solve it efficiently? Wish list: Expected time is polynomial in the number of constraints Expected time in terms of dimension isn t too bad.

3 The Linear Separability Problem Given a set X of points in R n, which is partitioned into two sets U and V, is there a plane that separates the points in U from the points in V? Figure: 2 variables, a whole bunch of restrictions!

Las Vegas Algorithms A Las Vegas Algorithm is a randomized algorithm that always gives correct results. Las Vegas algorithms either return a solution, or return failure. First introduced by Laszlo Babai in 1979 in the context of the graph isomorphism problem. Gamble with the quickness, but not with the result! Hence the name Las Vegas.

Las Vegas Algorithms A Las Vegas Algorithm is a randomized algorithm that always gives correct results. Las Vegas algorithms either return a solution, or return failure. First introduced by Laszlo Babai in 1979 in the context of the graph isomorphism problem. Gamble with the quickness, but not with the result! Hence the name Las Vegas. Are you feeling lucky?

6 Spoilers This paper explores a new algorithm for both linear and integer programs. For a linear problem with n constraints and d variables, the algorithm requires an expected arithmetic operation. 0(d 2 n) + (log n)o(d) d/2+o(1) + O(d 4 n log n) The paper also presents a related integer programming algorithm with expected O(2 d dn + 8 d n ln n ln n) + d O(d) φ ln n operations on numbers with d 0(1) φ bits, as n In both of these values, the constant factors do not depend on d or φ.

7 Linear Programming Algorithms The lay of the land for linear programming algorithms Meggido: O(2 2d n) Dyer and Frieze: Used random sampling to obtain expected O(d 3d n) time. This Paper: Leading term in n is O(d 2 n). Note the considerable improvement in d!

Lastly, this 2-norm mimimizer can be found quickly (Gill, Murray, Wright). 8 The Problem Determine the maximum x 1 coordinate of points satisfying all contraints, i.e. x 1 = max{x! Ax b}. Where A is an n d matrix. Each inequality in this system defines a closed halfspace H. Call the collection of these halfspaces S. Let P(S) be the polyhedron defined by S. It is no loss of generality to assume that the optimum x1 because we can take the x1 of minimum 2-norm. is unique, The x 1 of minimum 2-norm is unique, because of the convexity of the solution set P(S) {x x 1 = x 1 }.

9 Outline of Algorithm x s (S): a simplex like algorithm to solve LPs quickly. x r (S): A recursive algorithm designed to quickly throw away redundant constraints. xi (S): An iterative algorithm designed to weight the inequalities based on their importance. x m(s): A mixed algorithm which combines the first two algorithms. It calls x r (S), and then x i.

10 x r pseudocode Function x r {S : SetOfHalfspaces} Return x : LPoptimum V ; C d 9d 2 If n C d Then Return x s (S) Repeat choose R S \ V at random, R = r = d n x x r (R V ) V {H S x violates H} If V 2 n Then V V V Until V = Return x

11 x r details The main idea behind x r is the Helly s Theorem Theorem Helly Let S be a family of convex subsets of R d. If the intersection of every d + 1 of these sets is nonempty, then the whole collection has nonempty intersection. Applied to our context: The unique optimum of our contraints is determined by d or fewer constraints of S, which we call S. (Take the d + 1th constraint to be x 1 x 1 ). Idea: Randomly choose a subset of planes of a certain size, and optimize the objective over this subset. If this value violates few inequalities in S, the chosen subset is relevant.

x r details One issue with the x r algorithm is the recursive call. If we have a large amount of constraints to be discarded, it may take a long time to randomly sample the right ones. Each of these irrelevant random selections of R calls x s. These repeated, unnecessary calls to xs are the bottleneck that slows down xr. How can we encourage the algorithm to choose constraints that are relevant to S?

13 x i pseudocode Function x i {S : SetOfHalfspaces} Return x : LPoptimum for H S do w H 1; C d 9d 2 If n C d Then Return x s (S) Repeat choose R S at random, R = r = C d x x s (R) V {H S x violates H} If wv 2w(S)/(9d 1) Then for H V do w H 2w H od; Until V = Return x

x i details This reweighting technique allows us to choose constraints efficiently. Choose a random subset R, with probability of selecting a constraint proportional to their weight. Let V be the constraints violated by the optimum by that subset. If V is low, this subset of inequalities resembles S, so double the weights of these inqualities. Eventually, the constraints in S has large enough weight that the random subset R contains S. Question: Is doubling the weight optimum? Why not triple, or quadruple, the chosen weight? Would this make S R faster?

15 The Mixed Algorithm x m is the mixed algorithm, our white whale in terms of complexity. It is a version of the recursive algorithm xr that calls the iterative algorithm xi for its recursive calls. The motivation for this is to have a time bound with leading term O(d 2 n) while avoiding the large number of calls to xs of the recursive algorithm.

16 Complexity Analysis The success of all of these algorithms is based on the following lemma: Lemma In the recursive, iterative, and mixed algorithms, if the set V is nonempty, it contains a constraint of S. Draw a picture! This allows us to be sure that we are making progress with each iteration of the algorithm; elements of S are being reweighted.

Complexity Analysis How often is V 2 n? Lemma Let V S, and let R S \ V be a random subset of size r, with S \ V = n. Let V S be the set of constraints violated by x (R V ). Then the expected size of V is no more than d(n r + 1)/(r d). Which is used as a tool to prove Lemma The probability that any given execution of the loop body has V 2 n is at least 1/2, and so on average two executions are required to obtain a successful one.

18 Complexity Analysis Theorem Given an LP problem with b 0, the iterative algorithm x i requires O(d 2 log n) + (d log n)o(d) d/2+o(1) expected time, as n, where the constant factors do not depend on d. This is broken down into two steps The loop body is executed an expected O(d log n) times. The loop body requires O(dn) + O(d) d/2+o(1). This leading term in n, dn, is from the computation that determines V.

Theorem Complexity Analysis Algorithm x m requires O(d 2 n) + (d 2 log n)o(d) d/2+o(1) + O(d 4 n log n) expected time, as n, where the constant factors do not depend on d. Idea of proof: At most d + 1 successful ( V 2 n) iterations are needed so that S V. Take away: xi (S) does pretty well, but by paying really close attention to details we can reduce the leading term in n from n log n to n.

20 Integer Programming Algorithms Lenstra showed that integer linear programming can be solved in polynomial time when the number of variables is fixed. The fastest deterministic algorithm for this problem requires d O(d) nφ operations on d O(1) φ-bit numbers. Here φ is the facet complexity of the input, the maximum number of bits used to specify an input inequality constraint. Clarkson s algorithm in this paper appeals to Lenstra s algorithm.

21 The Integer Case The basis of the algorithm is a result analogous to Helly s Theorem. Theorem Doignon-Bell-Scarf There is a set S S with S 2 d 1 and with x (S) = x (S ). In other words, the optimimum of the ILP is determined by a small set.

22 The Integer Case The ILP algorithms are simply variations on the LP algorithms Sample sizes using 2 d rather than d Lenstra s algorithm in the base case The set S is not necessarily unique. Does Helly s Theorem guarantee uniqueness?

The Integer Case We have a similar lemma to the linear case Lemma Let V S, and let R S \ V be a random subset of size r > 2 d+1, with S \ V = n. Let V S be the set of constraints violated by x (R V ). Then with probability 1/2, V 2 d+1 n(ln r)/r.

The Integer Case Using this lemma, we have the following modifications for the ILP algorithms For the recursive algorithm, put the sample size at 2 d 2n ln n Use Lenstra s algorithm with n 2 2d+5 d With probability 1/2, the set V will contain no more than 2n ln n constraints. Require this for a successful iteration. For the iterative algorithm, use a sample size of 2 2d+4 (2d + 4), with a corresponding V bound of n(ln 2)/2 d+3.

25 The Integer Case Analysis of the ILP mixed algorithm is similar to the LP algorithm The top level does expected 2 d+1 n row operations, and generated 2 d+1 expected subproblems Each of these subproblems has no more than 2 d+1 2n ln n constraints. Each subproblem requires 2 d+1 ln n iterations. Each iteration requires 2 d+1 2n ln n row operations, and a call to Lenstra s algorithm.

These row operations are just the the evaluation of an input inequality at a given integral point. Clarkson claims that the integral points can be specified with 7d 3 φ bits. 26 The Integer Case These computations yield Theorem The ILP algorithm x m requires expected O(2 d n + 8 d n ln n ln n) row operations on O(d 3 φ)-bit vectors, and d O(d) φ ln n expected operations on O(d O(1) φ)-bit numbers, as n, where the constant factors do not depend on d or φ.

27 The Clarkson Soup Clarkson-Type Algorithms Input A limit on the number of restrictions necessary to characterize the problem (Helly s theorem, Doignon-Bell-Scarf theorem). A way to solve small problems quickly (Simplex-type algorithm, Lenstra s algorithm). A metric of how to verify how similar a subproblem is to the original. How many constraints does the solution to this subproblem violate? Output Improved complexity results!

The Clarkson Soup Theme of Clarkson Algorithms: Choose a boundary for the metric that guarantees forward progress will be made. I.E. If the number of original inequalities that are violated by the randomly selected constraints is small, then the subproblem contains (mostly) necessary constraints of S. Choose a sample size so that the probability of this success is greater than or equal to 1/2. Do a bunch of random sampling! Outsource small subproblems to whatever the best tool is on the market.

29 K-feasibility Clarkson Theorem DeLoera, Iskander, Louveaux If S is an integer program with k feasible points, there is a subproblem consisting of c(d, k) constraints with the same k-points (and no more). And so it begins! We want to use this theorem to establish a Clarkson-type algorithm to find the best k solutions to S. Given an arbitrary integer program, we can add the constraint objective S x k, where x k is the k-th best solution to turn satisfy the conditions for the theorem.

30 Looking Forward Open Questions and Ideas: How to check for k-feasibility quickly? Barvinok s algorithm counts the number of integral points in a polyhedron in polynomial time (in a fixed dimension). What kind of metric should be used? The number or original constraints violated by the solution to the subproblem will not work, since we aren t interested in the preservation of a single point, but k of them. Also, we don t have these k points in our hand like we do for Lensta s and the simplex algorithm!

Violator Spaces One promising generalization of Clarkson-like-algorithms is Violator Spaces. A Violator Space is a pair (H, V ), where H is a finite set and V is a mapping 2 H 2 H such that Consistency: G V (G) = holds for all G H Locality: For all F G H, where G V (F ) =. We think of H as the set of constraints that come with a problem, and the map V is the map that, given a subset G returns the constraints in H that are violated by the optimum over G. (Picture)

Basis and Combinatorial Dimension We say that B H is a basis if for all proper subsets F B we have B V (F ). In other words, the smallest (by inclusion) set of constraints that have the same optimum objective value. Let (H, V ) be a violator space. The size of a largest basis is called the combinatorial dimension of (H, V ).

Basis and Combinatorial Dimension We say that B H is a basis if for all proper subsets F B we have B V (F ). In other words, the smallest (by inclusion) set of constraints that have the same optimum objective value. Let (H, V ) be a violator space. The size of a largest basis is called the combinatorial dimension of (H, V ). Helly s Theorem Anybody?! It turns out that Clarkson s Algorithm holds for generalized violator spaces. (After proving a non-trivial Sampling Lemma.)

34 Conclusion Thanks for listening!