mixed-integer convex optimization

Similar documents
Mixed-integer convex optimization: outer approximation algorithms and modeling power. Miles Lubin

A Lifted Linear Programming Branch-and-Bound Algorithm for Mixed Integer Conic Quadratic Programs

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg

LaGO - A solver for mixed integer nonlinear programming

The Supporting Hyperplane Optimization Toolkit A Polyhedral Outer Approximation Based Convex MINLP Solver Utilizing a Single Branching Tree Approach

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs

A NEW SEQUENTIAL CUTTING PLANE ALGORITHM FOR SOLVING MIXED INTEGER NONLINEAR PROGRAMMING PROBLEMS

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs

Conic Optimization via Operator Splitting and Homogeneous Self-Dual Embedding

13. Cones and semidefinite constraints

Embedding Formulations, Complexity and Representability for Unions of Convex Sets

A Center-Cut Algorithm for Quickly Obtaining Feasible Solutions and Solving Convex MINLP Problems

Review of Mixed-Integer Nonlinear and Generalized Disjunctive Programming Methods

The AIMMS Outer Approximation Algorithm for MINLP

The AIMMS Outer Approximation Algorithm for MINLP

Introduction to Mathematical Programming IE406. Lecture 20. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 1. Dr. Ted Ralphs

Mathematical Programming and Research Methods (Part II)

Conic Duality. yyye

Convex Optimization M2

Penalty Alternating Direction Methods for Mixed- Integer Optimization: A New View on Feasibility Pumps

Linear programming and duality theory

From the Separation to the Intersection Sub-problem in Benders Decomposition Models with Prohibitively-Many Constraints

CS 473: Algorithms. Ruta Mehta. Spring University of Illinois, Urbana-Champaign. Ruta (UIUC) CS473 1 Spring / 36

LECTURE 18 LECTURE OUTLINE

Integer Programming Chapter 9

Financial Optimization ISE 347/447. Lecture 13. Dr. Ted Ralphs

Integer Programming Theory

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Stochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition

ALGORITHMS AND SOFTWARE FOR CONVEX MIXED INTEGER NONLINEAR PROGRAMS

An extended supporting hyperplane algorithm for convex MINLP problems

Shiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization

On the Global Solution of Linear Programs with Linear Complementarity Constraints

COMS 4771 Support Vector Machines. Nakul Verma

Exploiting Low-Rank Structure in Semidenite Programming by Approximate Operator Splitting

Optimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.

From the Separation to the Intersection Sub-problem in Benders Decomposition Models with Prohibitively-Many Constraints

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

Cutting Planes for Some Nonconvex Combinatorial Optimization Problems

Week 5. Convex Optimization

Optimality certificates for convex minimization and Helly numbers

Approximation Algorithms

Tutorial on Convex Optimization for Engineers

Strategies for Using Algebraic Modeling Languages to Formulate Second-Order Cone Programs

Applied Lagrange Duality for Constrained Optimization

arxiv: v4 [math.oc] 4 Nov 2018

Cloud Branching MIP workshop, Ohio State University, 23/Jul/2014

Linear Programming Duality and Algorithms

Convex Optimization: from Real-Time Embedded to Large-Scale Distributed

Lecture 2. Topology of Sets in R n. August 27, 2008

A Nonlinear Presolve Algorithm in AIMMS

2. Convex sets. x 1. x 2. affine set: contains the line through any two distinct points in the set

MathOptInterface and JuMP Miles Lubin Google JuMP-dev 2018

Research Interests Optimization:

Math 5593 Linear Programming Lecture Notes

6 Randomized rounding of semidefinite programs

CS 372: Computational Geometry Lecture 10 Linear Programming in Fixed Dimension

Convex Optimization MLSS 2015

Linear Programming in Small Dimensions

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

Lecture 5: Duality Theory

Foundations of Computing

Optimality certificates for convex minimization and Helly numbers

Convex Optimization - Chapter 1-2. Xiangru Lian August 28, 2015

25. NLP algorithms. ˆ Overview. ˆ Local methods. ˆ Constrained optimization. ˆ Global methods. ˆ Black-box methods.

What's New In Gurobi 5.0. Dr. Edward Rothberg

Algorithms for Decision Support. Integer linear programming models

/ Approximation Algorithms Lecturer: Michael Dinitz Topic: Linear Programming Date: 2/24/15 Scribe: Runze Tang

2. Convex sets. affine and convex sets. some important examples. operations that preserve convexity. generalized inequalities

A Brief Overview of Optimization Problems. Steven G. Johnson MIT course , Fall 2008

Robust Optimization in AIMMS

CMU-Q Lecture 9: Optimization II: Constrained,Unconstrained Optimization Convex optimization. Teacher: Gianni A. Di Caro

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS

Introduction to Convex Optimization. Prof. Daniel P. Palomar

CME307/MS&E311 Theory Summary

Modern Benders (in a nutshell)

Decomposition of loosely coupled integer programs: A multiobjective perspective

Lecture 12: Feasible direction methods

Submodularity Reading Group. Matroid Polytopes, Polymatroid. M. Pawan Kumar

Investigating Mixed-Integer Hulls using a MIP-Solver

Lagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Lagrangian Relaxation: An overview

Convex Optimization. Convex Sets. ENSAE: Optimisation 1/24

CS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm. Instructor: Shaddin Dughmi

Benders in a nutshell Matteo Fischetti, University of Padova

Global Solution of Mixed-Integer Dynamic Optimization Problems

Linear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions

CS 435, 2018 Lecture 2, Date: 1 March 2018 Instructor: Nisheeth Vishnoi. Convex Programming and Efficiency

11 Linear Programming

Convex Optimization and Machine Learning

Mixed Integer Second Order Cone Optimization (MISOCO): Cuts, Warm Start, and Rounding

60 2 Convex sets. {x a T x b} {x ã T x b}

CS675: Convex and Combinatorial Optimization Spring 2018 The Simplex Algorithm. Instructor: Shaddin Dughmi

LP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008

LECTURE NOTES Non-Linear Programming

Selected Topics in Column Generation

The Gurobi Optimizer. Bob Bixby

Transcription:

mixed-integer convex optimization Miles Lubin with Emre Yamangil, Russell Bent, Juan Pablo Vielma, Chris Coey April 1, 2016 MIT & Los Alamos National Laboratory

First, Mixed-integer linear programming (MILP), min x c T x subject to Ax = b, x 0, x i Z, i I - Despite NP-Hardness, many problems of practical interest can be solved to optimality or near optimality - Algorithms are based on LP relaxations, branch & bound, cuts, preprocessing, heuristics,... - 50+ years of commercial investment in developing these techniques 1

who uses milp? http://www.gurobi.com/company/example-customers 2

example: unit commitment Minimize unit-production costs + fixed operating costs subject to: - Total generation = total demand, hourly over 24h - Generation and transmission limits If a generator is on, it produces within some interval [l, u], otherwise its production is zero - Linear approximation of nonlinear powerflow laws 1% reduction in generation gives billions of dollars per year cost reduction. Worth spending time to improve optimality. 3

example: sparse linear regression min x Ax b 2 subject to x 0 k, Suppose we know optimal solution satisfies x M. We have an equivalent mixed-integer quadratic formulation min x,y subject to Ax b 2 yi k, x i My i x i My i y {0, 1} n 4

example: sparse linear regression min x Ax b 2 subject to x 0 k, These are solvable up to 1000s of dimensions using modern commercial solvers (Bertsimas and King, 2016). 5

example: experimental design We take m measurements y i by performing a set of experiments a i, y i = a T i x + ϵ i with noise ϵ i independent standard Gaussian. Standard confidence set for x R n has form {z (z ˆx) T E 1 (z ˆx) β} where ˆx is the maximum likelihood estimate of x, ( m ) 1 E = a i a T i, i=1 and β depends on n confidence level α. 6

example: experimental design Say we want to choose experiments a i in an optimal way from a set of potential experiments v 1, v 2,..., v p. Introduce an integer variable m j counting how many times we perform experiment v j. Then p E = m j v j v T j We could choose to minimize the volume of the confidence ellipsoid, j=1 1. max m subject to det(e 1 ) mj = k, m 0, m Z n. 7

example: experimental design max m subject to det(e 1 ) mj = k, m 0, m Z n. becomes max m subject to log(det(e 1 )) mj = k, m 0, m Z n. - Convex optimization problem when we relax integrality - Boyd (Ch. 7.5) proposes to solve convex relaxation and round, but maybe we can solve the original problem to optimality! 8

Mixed-integer convex programming (MICP), min x f(x) subject to g j (x) 0, j x i Z, i I Objective function f and constraints g j convex 9

what can be modeled using micp? Disjunctions/unions of compact convex sets (Stubbs and Mehrotra, 1999) {x : f 1 (x) 0 or f 2 (x) 0} 3 2 1 0-1 -2-3 -3-2 -1 0 1 2 3 10

3 2 1 0-1 -2-3 -3-2 -1 0 1 2 3 x belongs to the convex hull iff u 1, u 2, λ 1, λ 2 such that x = λ 1 u 1 + λ 2 u 2 λ 1 + λ 2 = 1 λ 1 0, λ 2 0 f 1 (u 1 ) 0 f 2 (u 2 ) 0 11

x = λ 1 u 1 + λ 2 u 2 λ 1 + λ 2 = 1 λ 1 0, λ 2 0 f 1 (u 1 ) 0 f 2 (u 2 ) 0 - If we impose λ 1, λ 2 {0, 1} then have conditions for the union, but not convex 12

x = λ 1 u 1 + λ 2 u 2 λ 1 + λ 2 = 1 λ 1 0, λ 2 0 f 1 (u 1 ) 0 f 2 (u 2 ) 0 - If we impose λ 1, λ 2 {0, 1} then have conditions for the union, but not convex - Instead, introduce z 1 = λ 1 u 1, z 2 = λ 2 u 2 12

x = z 1 + z 2 λ 1 + λ 2 = 1 λ 1 0, λ 2 0 f 1 (z 1 /λ 1 ) 0 f 2 (z 2 /λ 2 ) 0 - Still not convex 13

x = z 1 + z 2 λ 1 + λ 2 = 1 λ 1 0, λ 2 0 λ 1 f 1 (z 1 /λ 1 ) 0 λ 2 f 2 (z 2 /λ 2 ) 0 λ 1, λ 2 {0, 1} - Perspective function of any convex function is convex. 14

- Which nonconvex sets can be modeled with MICP? Finite unions of compact, convex sets Conjecture: the following set is not MICP representable 3 2 1 0-1 -2-3 -3-2 -1 0 1 2 3 15

State of the art for solving MICP min x subject to f(x) g j (x) 0, j x i Z, i I - f, g j convex, smooth - Bonmin, KNITRO, α-ecp, DICOPT, FilMINT, MINLP_BB, SBB,... 16

outer approximation algorithm Idea: Solve an alternating sequence of MILP and convex problems using existing, good solvers. Given points x 1,..., x R, solve mixed-integer linear relaxation: min x t subject to g j (x r ) + g j (x r ) T (x x r ) 0, j, r = 1,..., R f(x r ) + f(x r ) T (x x r ) t, x i Z, i I r = 1,..., R Then solve convex problem with integer values fixed to get a new feasible solution, and repeat with R = R + 1. 17

not good enough... [...] solution algorithms for [MICP] have benefit from the technological progress made in solving MILP and NLP. However, in the realm of [MICP], the progress has been far more modest, and the dimension of solvable [MICP] by current solvers is small when compared to MILPs and NLPs. Bonami, Kilinç, and Linderoth (2012) 18

what can go wrong? Classical approaches form polyhedral outer approximations by a finite collection of gradient linearizations. But a good polyhedral outer approximation might need too many linear constraints. Recall: { B 1 := {x : x 1 1} = x : } n s i x i 1, s { 1, +1} n i=1 19

B 1 := {x : x 1 1} = { x : } n s i x i 1, s { 1, +1} n Standard trick in linear programming is to introduce extra variables: { B 1 = x R n : y R n s.t. x y, x y, } y i 1 i The new formulation has 2n variables and 2n + 1 constraints versus n variables and 2 n constraints. Not bad. i=1 We call this an extended formulation. 20

the same happens with smooth constraints Hijazi et al. (2014): B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 21

what can go wrong? Hijazi et al. (2014): B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 - B n is empty - No outer approximating hyperplane can exclude more than one integer lattice point - So any polyhedral outer approximation which contains no integer vertices needs at least 2 n hyperplanes 22

how to fix it? B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 Consider the equivalent formulation in an extended set of variables: n ˆB n := x {0, 1}n, z R n : z j n 1 ( 4, x j 2) 1 2 z j j - B n = proj x ˆB n j=1 23

how to fix it? B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 Consider the equivalent formulation in an extended set of variables: n ˆB n := x {0, 1}n, z R n : z j n 1 ( 4, x j 2) 1 2 z j j - B n = proj x ˆB n j=1 - If you form a polyhedral relaxation of each ( x j 1 2) 2 zj individually, then 2n hyperplanes in R 2n is sufficient to exclude all integer points 23

how to fix it? B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 Consider the equivalent formulation in an extended set of variables: n ˆB n := x {0, 1}n, z R n : z j n 1 ( 4, x j 2) 1 2 z j j - B n = proj x ˆB n j=1 - If you form a polyhedral relaxation of each ( x j 1 2) 2 zj individually, then 2n hyperplanes in R 2n is sufficient to exclude all integer points - Outer approximation algorithm now converges in 2 iterations 23

how to fix it? B n := x {0, 1}n : n ( x j 1 ) 2 n 1 2 4 j=1 Consider the equivalent formulation in an extended set of variables: n ˆB n := x {0, 1}n, z R n : z j n 1 ( 4, x j 2) 1 2 z j j - B n = proj x ˆB n j=1 - If you form a polyhedral relaxation of each ( x j 1 2) 2 zj individually, then 2n hyperplanes in R 2n is sufficient to exclude all integer points - Outer approximation algorithm now converges in 2 iterations - Why? Small polyhedron in higher dimension can have exponentially many facets in lower dimension 23

extended formulation for second-order cone SOC n := {(t, x) R n : x 2 t} Second order cone programming (SOCP) generalizes convex quadratic programming (QP). 24

extended formulation for second-order cone SOC n := {(t, x) R n : x 2 t} Second order cone programming (SOCP) generalizes convex quadratic programming (QP). Vielma, Dunning, Huchette, Lubin (2015): - For MISOCP: Rewrite as x 2 i t 2 i z i t, where z i x 2 i /t i - Implemented by CPLEX and Gurobi within months of publication 24

we know what to do... Hijazi et al. (2014) propose to reformulate constraints of the form to n g i (x i ) 0 i=1 n z i 0 and g i (x i ) z i i = 1,..., n, i=1 where g i univariate, smooth, convex. - Very impressive computational results - Paper submitted in 2011, but there s still no solver which automates this transformation 25

Why has nobody implemented this? 26

Why has nobody implemented this? min x f(x) subject to g j (x) 0, j x i Z, i I - MICP solvers view f and g j through black-box oracles for evaluations of values and derivatives - Summation structure not accessible through the black box 26

Mixed-integer disciplined convex programming (MIDCP), min x f(x) subject to g j (x) 0, j x i Z, i I Objective function f and constraints g j disciplined convex Forget about derivatives 27

Models with integer and binary variables must still obey all of the same disciplined convex programming rules that CVX enforces for continuous models. For this reason, we are calling these models mixed-integer disciplined convex programs or MIDCPs. 28

What is a disciplined convex function? (Grant, Boyd, Ye, 2006) - Expressed as composition of basic atoms. Rules of composition prove convexity. - Example x 2 + y 2 versus [x, y] 2, xy versus geomean(x, y). Why disciplined convex? - Functions expressed in this form can be automatically converted to conic optimization problems. We have good solvers and theory for conic optimization. 29

conic optimization min x c T x s. t. Ax = b x K - K is a cone if x K implies αx K α 0 - Ususally we consider K = K 1 K s where each K j is one of a small number of cones like R n +, SOC n, the cone of positive semidefinite matrices... 30

min x DCP f(x) s. t. g j (x) 0, j CVX CVXPY Convex.jl min x Conic c T x s. t. Ax = b x K ECOS SCS Mosek Optimal solution x 31

Example: exp(c T i x + b i) t i is expressed in conic form as z i t and (c T i x + b i, 1, z i ) EXP. i where EXP = cl{(x, y, z) R 3 : y exp(x/y) z, y > 0} is the exponential cone. 32

Example: exp(c T i x + b i) t i is expressed in conic form as z i t and (c T i x + b i, 1, z i ) EXP. i where EXP = cl{(x, y, z) R 3 : y exp(x/y) z, y > 0} is the exponential cone. The translation to conic form already produces a formulation with additional variables ideal for polyhedral outer approximation. 32

Example: exp(x 2 + y 2 ) t is DCP compliant. CVX will generate equivalent formulation exp(z) t where x 2 + y 2 z which is then translated to EXP and second-order cones. - Tawarmalani and Sahinidis (2005) show that polyhedral outer approximations that exploit composition structure have extra strength. This comes for free with conic representations 33

- Mixed-integer second order cone programming (MISOCP) has rapidly improving commercial support by Gurobi/CPLEX Gurobi claims 2.8x improvement from version 6.0 to 6.5 (1 year) - Otherwise (exponential cones, SDP), mainly research codes (ecos_bb by Han Wang, 2014; SCIP-SDP by Mars, 2012) - For MISOCP, no established open-source codes 34

How general is MISOCP? MINLPLIB2 benchmark library designed for derivative-based mixed-integer nonlinear solvers. We classified all of the convex instances by conic representability. Of the 333 convex instances, 65% are MISOCP representable. - tls6 instance previously unsolved. We translated it to Convex.jl and it was solved by Gurobi! 35

What about the remaining 35% of problems? Of remaining 116 instances, in 107 the only nonlinear expressions involve exp and log. 36

What about the remaining 35% of problems? Of remaining 116 instances, in 107 the only nonlinear expressions involve exp and log. Recall the exponential cone: EXP = cl{(x, y, z) R 3 : y exp(x/y) z, y > 0} - Closed, convex, nonsymmetric cone. Theoretically tractable and supported by ECOS (Akle, 2015) 36

What about the remaining 9 of 333 instances? - Seven representable by combination of EXP and second-order cones (SOC) - Two representable by the power cone POW α = {(x, y, z) R 3 : z x a y 1 a, x 0, y 0} 37

Folklore: Almost all convex optimization problems of practical interest can be represented as conic programming problems using second-order, positive semidefinite, exponential, and power cones. 38

So cones are general and provide a good extended formulation of the original convex problem. How do we apply polyhedral outer approximation to mixed-integer conic problems? - We ll state the first finite-time outer approximation algorithm for general mixed-integer conic optimization 39

min x MIDCP f(x) s. t. g j (x) 0, j x i Z, i I CVX CVXPY Convex.jl min x MICONE c T x s. t. Ax = b x K x i Z, i I??? Optimal solution x 40

mixed-integer conic programming min x,z c T z s.t. A x x + A z z = b (MICONE) L x U, x Z n, z K, - K = K 1 K 2 K r, where each K i is a standard cone: nonnegative orthant, second-order cone, exponential cone, power cone, or even positive semidefinite cone - WLOG, integer variables do not belong to cones, have zero objective coefficients 41

Need to outer approximate cone K with finite number of linear constraints... 42

Need to outer approximate cone K with finite number of linear constraints... Given a closed, convex cone K, the dual cone K is the set such that z K iff z T β 0 β K. - Nonnegative orthant, second-order and positive semidefinite cone are self dual. Exponential and power cones are not - Any finite subset of K yields a polyhedral outer approximation of K! 42

min x,z c T z s.t. A x x + A z z = b (MIOA(T)) L x U, x Z n, β T z 0 β T, - If T = K, then equivalent to MICONE - If T K and T <, then polyhedral outer approximation MILP relaxation of MICONE - Claim: T K, T < such that MIOA(T) is equivalent to MICONE 43

Consider the subproblem with integer values x = ˆx fixed: The dual of CP(ˆx) is vˆx = min z c T z s.t. A z z = b A xˆx, (CP(ˆx)) z K. max β,λ s.t. λ T (b A xˆx) β = c A T zλ β K. 44

the conic outer approximation algorithm - Start with T = (or small) - Until the optimal objective of MIOA(T) is the optimal objective of the best feasible solution: Let ˆx be the optimal integer solution of the MILP problem MIOA(T) Solve conic problem CP(ˆx), add dual solution βˆx to T If CP(ˆx) was feasible, record new feasible solution to MICONE 45

the conic outer approximation algorithm - Start with T = (or small) - Until the optimal objective of MIOA(T) is the optimal objective of the best feasible solution: Let ˆx be the optimal integer solution of the MILP problem MIOA(T) Solve conic problem CP(ˆx), add dual solution βˆx to T If CP(ˆx) was feasible, record new feasible solution to MICONE Finite termination assuming strong duality holds at subproblems 45

the conic outer approximation algorithm - Easy to implement, requires black box MILP and conic solvers - Very often, number of iterations is < 10, sometimes 10-30. Worst observed is 171. - Each MILP subproblem could be NP-Hard, but takes advantage of fast solvers 46

vˆx = min z c T z max β,λ λ T (b A xˆx) s.t. A z z = b A xˆx, s.t. β = c A T zλ z K. β K. Suppose CP(ˆx) is feasible and strong duality holds at the optimal primal-dual solution (zˆx, βˆx, λˆx ). Then for any z with A z z = b A xˆx and β Ṱ x z 0, we have ct z vˆx. 47

vˆx = min z c T z max β,λ λ T (b A xˆx) s.t. A z z = b A xˆx, s.t. β = c A T zλ z K. β K. Suppose CP(ˆx) is feasible and strong duality holds at the optimal primal-dual solution (zˆx, βˆx, λˆx ). Then for any z with A z z = b A xˆx and β Ṱ x z 0, we have ct z vˆx. Including β Ṱ x z 0 in polyhedral relaxation implies that if x = ˆx is the optimal (integer part of the) solution of the relaxation, then the objective value of the relaxation is at least vˆx. Proof. β Ṱ x z = (c AT zλˆx ) T z = c T z λ Ṱ x (b A xˆx) = c T z vˆx 0. 47

vˆx = min z c T z max β,λ λ T (b A xˆx) s.t. A z z = b A xˆx, s.t. β = c A T zλ z K. β K. Given ˆx, assume CP(ˆx) is infeasible and its dual is unbounded, such that we have a ray (βˆx, λˆx ) satisfying βˆx K, β = A T zλˆx, and λ Ṱ x (b A xˆx) > 0. Then for any z satisfying A z z = b A xˆx we have β Ṱ x z < 0. Including β Ṱ x z 0 in polyhedral relaxation implies x = ˆx is infeasible to the relaxation. 48

vˆx = min z c T z max β,λ λ T (b A xˆx) s.t. A z z = b A xˆx, s.t. β = c A T zλ z K. β K. Given ˆx, assume CP(ˆx) is infeasible and its dual is unbounded, such that we have a ray (βˆx, λˆx ) satisfying βˆx K, β = A T zλˆx, and λ Ṱ x (b A xˆx) > 0. Then for any z satisfying A z z = b A xˆx we have β Ṱ x z < 0. Including β Ṱ x z 0 in polyhedral relaxation implies x = ˆx is infeasible to the relaxation. Proof. β Ṱ x z = λṱ x A zz = λ Ṱ x (b A xˆx) < 0. 48

Claim: T K, T < such that MIOA(T) is equivalent to MICONE 49

Claim: T K, T < such that MIOA(T) is equivalent to MICONE Proof. Solve CP(ˆx) for all possible ˆx (finite). Then include all corresponding dual vectors βˆx in T. If x is the integer solution of MIOA(T), then CP(x ) is feasible from the second lemma. From first lemma, objective of MIOA(T) is at least the optimal value of CP(x ), but it s also a relaxation so x must be optimal. 49

recap - Polyhedral approximations can be made much better by introducing extra variables 50

recap - Polyhedral approximations can be made much better by introducing extra variables - Summation structure and compositions of convex functions should be broken up 50

recap - Polyhedral approximations can be made much better by introducing extra variables - Summation structure and compositions of convex functions should be broken up - Nobody automated this before! (Derivative-based view prevents implementation) 50

recap - Polyhedral approximations can be made much better by introducing extra variables - Summation structure and compositions of convex functions should be broken up - Nobody automated this before! (Derivative-based view prevents implementation) - DCP to the rescue. Conic form is general and encodes the information we need as a solver 50

recap - Polyhedral approximations can be made much better by introducing extra variables - Summation structure and compositions of convex functions should be broken up - Nobody automated this before! (Derivative-based view prevents implementation) - DCP to the rescue. Conic form is general and encodes the information we need as a solver - Proposed first outer approximation algorithm for MICONE, based on conic duality 50

Computational results 51

Pajarito New solver! - 1000 lines of Julia - Input is in conic form, accessible through Convex.jl DCP language - Works with any MILP and conic solvers supported in Julia - For tough conic problems, also supports traditional nonlinear solvers - Released last week! 52

Translated 194 convex instances from MINLPLIB2 from AMPL into Convex.jl (heroic parsing work by E. Yamangil). Compare with Bonmin s outer approximation algorithm using CPLEX as the MILP solver. We use the same version of CPLEX plus KNITRO as NLP (conic) solver. 53

(a) Solution time (b) Number of OA iterations Figure: Comparison performance profiles for SOC representable instances (Bonmin >30 sec) 54

(a) Solution time (b) Number of OA iterations Figure: Comparison performance profiles for SOC representable instances (Bonmin >30 sec) - Dominates on iteration count. Haven t optimized Pajarito so Bonmin is faster on the easier problems. 54

(a) Solution time (b) Number of OA iterations Figure: Comparison performance profiles for SOC representable instances (Bonmin >30 sec) - Dominates on iteration count. Haven t optimized Pajarito so Bonmin is faster on the easier problems. - CPLEX is usually the fastest on MISOCPs. Use that instead. 54

Almost none of the non-misocp instances in MINLPLIB2 are hard! 55

gams01 gams01 unsolved instance which is EXP+SOC representable - 145 variables. 1268 constraints (1158 are linear) - Previous best bound: 1735.06. Best known solution: 21516.83 - Pajarito solved in 6 iterations (< 10 hours). Optimal solution is 21380.20. 56

ongoing and future work - Extensions to geometric programming and semidefinite programming - What happens when strong duality fails? - Lots of room for improvement in reliability of conic solvers (especially exponential cone) 57

Pajarito Mountain, New Mexico 58

Thanks! Questions? https://github.com/mlubin/pajarito.jl http://arxiv.org/abs/1511.06710 59