Bilinear Programming
|
|
- June Lane
- 6 years ago
- Views:
Transcription
1 Bilinear Programming Artyom G. Nahapetyan Center for Applied Optimization Industrial and Systems Engineering Department University of Florida Gainesville, Florida address: 1 Introduction A function f(x, y) is called bilinear if it reduces to a linear one by fixing the vector x or y to a particular value. In general, a bilinear function can be represented as follows: f(x, y) = a T x + x T Qy + b T y, where a, x R n, b, y R m, and Q is a matrix of dimension n m. It is easy to see that bilinear functions compose a subclass of quadratic functions. We refer to optimization problems with bilinear objective and/or constraints as bilinear problems, and they can be viewed as a subclass of quadratic programming. Bilinear programming has various applications in constrained bimatrix games, Markovian assignment and complementarity problems. Many 0-1 integer programs can be formulated as bilinear problems. An extensive discussion of different applications can be found in [3]. Concave piecewise linear and fixed charge network flow problems, which are very common in the supply chain management, can be also solved using bilinear formulations (see, e.g., [5] and [6]). 2 Formulation Despite a variety of different bilinear problems, most of the practical problems involve a bilinear objective function and linear constraints, and theoretical results are derived for those cases. In our discussion we consider the following bilinear problem, which we refer to as BP. min x X,y Y f(x, y) = at x + x T Qy + b T y, 1
2 where X and Y are nonempty polyhedra. The BP formulation is also known as a bilinear problem with a disjoint feasible region because the feasibility of x (y) is independent form the choice of the vector y (x). 2.1 Equivalence to Other Problems Below we discuss some theoretical results, which reveal the equivalence between bilinear problems and some of concave minimization problems. Let V (x) and V (y) denote the set of vertices of X and Y, respectively, and g(x) = min y Y f(x, y) = a T x + min y Y {x T Qy + b T y}. Note that min y Y f(x, y) is a linear programm. Because the solution of a linear problem attains on a vertex of the feasible region, g(x) = min y Y f(x, y) = min y V (Y ) f(x, y). Using those notations, the BP problem can be restated as min f(x, y) = min {min x X,y Y x X y Y f(x, y)} = min x X { min y V (Y ) f(x, y)} = min g(x). (1) x X Observe that the set of vertices of Y is finite, and for each y Y, f(x, y) is a linear function of x; therefore, function g(x) is a piecewise linear concave function of x. From the later it follows that BP is equivalent to a piecewise linear concave minimization problem with linear constraints. It also can be shown that any concave minimization problem with a piecewise linear separable objective function can be reduced to a bilinear problem. To establish this relationship consider the following optimization problem: min x X φ i (x i ), (2) i where X is an arbitrary nonempty set of feasible vectors, and φ i (x i ) is a concave piecewise linear function of only one component x i, i.e., c 1 i x i + s 1 i (= φ 1 i (x i )) x i [λ 0 i, λ 1 i ) c φ i (x i ) = 2 i x i + s 2 i (= φ 2 i (x i )) x i [λ 1 i, λ 2 i ), c n i i x i + s n i i (= φ n i i (x i )) x i [λ n i 1 i with c 1 i > c 2 i > > c n i i. Let K i = {1, 2,..., n i }. Because of the concavity of φ i (x i ), the function can be written in the following alternative form, λ n i φ i (x i ) = min k K i {φ k i (x i )} = min k K i {c k i x i + s k i }. (3) Construct the following bilinear problem: min f(x, y) = φ k i (x i )yi k = (c k i x i + s k i )yi k (4) x X,y Y i k K i i k K i where Y = [0, 1] i K i. The proof of the following theorem follows directly from equation (3), and for details we refer to the paper [5]. 2 i ]
3 Theorem 1 If (x, y ) is a solution of the problem (4) then x is a solution of the problem (2). Observe that X is not required to be a polytop. If X is a polytop then the structure of the problem (4) is similar to BP. Furthermore, it can be shown that any quadratic concave minimization problem can be reduced to a bilinear problem. Specifically, consider the following optimization problem: min φ(x) = x X 2aT x + x T Qx, (5) where Q is a symmetric negative semi-definite matrix. Construct the following bilinear problem min f(x, y) = x X,y Y at x + a T y + x T Qy, (6) where Y = X. Theorem 2 (see [2]) If x is a solution of the problem (5) then (x, x ) is a solution of the problem (6). If (ˆx, ŷ) is a solution of the problem (6) then ˆx and ŷ solve the problem (5). 2.2 Properties of a Solution In the previous section we have shown that BP is equivalent to a piecewise linear concave minimization problem. On the other hand it is well known that a concave minimization problem over a polytop attains its solution on a vertex (see, for instance, [1]). The following theorem follows from this observation. Theorem 3 (see [2] and [1]) If X and Y are bounded then there is an optimal solution of BP, (x, y ), such that x V (X) and y V (Y ). Let (x, y ) denote a solution of BP. By fixing the vector x to the value of the vector x, the BP problem reduces to a linear one, and y should be a solution of the resulting problem. From the symmetry of the problem, a similar result holds by fixing the vector y to the value of the vector y. The following theorem is a necessary optimality condition, and it is a direct consequence of the above discussion. Theorem 4 (see [2] and [1]) If (x, y ) is a solution of the BP problem, then min x X f(x, y ) = f(x, y ) = min y Y f(x, y) (7) However, (7) is not a sufficient condition. In fact it can only guarantee a local optimality of (x, y ) under some additional requirements. In particular, y has to be the unique solution of min y Y f(x, y) problem. From the later it follows that f(x, y ) < f(x, y), y V (Y ), y y. Because of the continuity of the function f(x, y), for any y V (y), y y, f(x, y ) < f(x, y) in a small neighborhood U y of the point x. Let U = y V (Y ),y y U y. 3
4 Then f(x, y ) < f(x, y), x U, y V (Y ), y y. At last observe that Y is a polytop, and any point of the set can be expressed through a convex combination of its vertices. From the later it follows that f(x, y ) f(x, y), x U, y Y, which completes the proof of the following theorem. Theorem 5 If (x, y ) satisfies the condition (7) and y is the unique solution of the problem min y Y f(x, y) then (x, y ) is a local optimum of BP. Recall that BP is equivalent to a piecewise concave minimization problem. Under the assumptions of the theorem, it is easy to show that x is a local minimum of the function g(x) as well (see [2]). 3 Methods In this section we discuss methods to find a solution of a bilinear problem. Because BP is equivalent to a piecewise linear concave minimization problem, any solution algorithm for the later can be used to solve the former. In particular, one can employ a cutting plain algorithms developed for those problems. However, the symmetric structure of the BP problem allows constructing more efficient cuts. In the paper [4], the author discusses an algorithm, which converges to a solution that satisfies condition (7), and then proposes a cutting plain algorithm to find the global minimum of the problem. Assume that X and Y are bounded. Algorithm 1, which is also known as the mountain climbing procedure, starts from an initial feasible vector y 0 and iteratively solves two linear problems. The first LP is obtained by fixing the vector y to the value of the vector y m 1. The solution of the problem is used to fix the value of the vector x and construct the second LP. If f(x m, y m 1 ) f(x m, y m ), then we continue solving the linear problems by fixing the vector y to the value of y m. If the stopping criteria is satisfied, then it is easy to show that the vector (x m, y m ) satisfies the condition (7). In addition, observe that V (X) and V (Y ) are finite. From the later and the fact that f(x m, y m 1 ) f(x m, y m ) it follows that the algorithm converges in a finite number of iterations. Let (x, y ) denote the solution obtained by the Algorithm 1. Assuming that the vertex x is not degenerate, denote by D the set of directions d j along the ages emanating from the point x. Recall that g(x) = min y Y f(x, y) is a concave function. To construct a valid cut, for each direction d j find the maximum value of θ j such that g(x + θ j d j ) f(x, y ) ε, i.e., θ j = argmax{θ j g(x + θ j d j ) f(x, y ) ε}, Algorithm 1 : Mountain Climbing Procedure Step 1: Let y 0 Y denote an initial feasible solution, and m 1. Step 2: Let x m = argmin x X {f(x, y m 1 )}, and y m = argmin y Y {f(x m, y)}. Step 3: If f(x m, y m 1 ) = f(x m, y m ) then stop. Otherwise, m m + 1 and go to Step 2. 4
5 Algorithm 2 : Cutting Plane Algorithm Step 1: Apply Algorithm 1 to find a vector (x, y ) that satisfies the relationship (7). Step 2: Based on the solution (x, y ), compute the appropriate cuts and construct the sets X 1 and Y 1. Step 3: If X 1 = or Y 1 =, then stop; (x, y ) is a global ε-optimal solution. Otherwise, X X 1, Y Y 1, and go to Step 1. where ε is a small positive number. Let C = (d 1,..., d n ), { ( 1 1 x = x,..., 1 ) T C 1 (x x ) 1}, θ 1 θ n and X 1 = X 1 x. If X 1 = then min f(x, y) x X,y Y f(x, y ) ε, and (x, y ) is a global ε-optimum of the problem. If X 1 then one can replace X by the set X 1, i.e, consider the optimization problem min f(x, y), x X 1,y Y and run Algorithm 1 to find a better solution. However, because of the symmetric structure of the problem, a similar procedure can be applied to construct a cut for the set Y. Let 1 y denote the corresponding half-space, and Y 1 = Y 1 y. By updating both sets, i.e., considering the optimization problem min f(x, y), x X 1,y Y 1 the cutting plane algorithm (see Algorithm 2) might find a global solution of the problem using less number of iterations. References [1] Horst R, Pardalos P, Thoai N. Introduction to global optimization, 2-nd Edition, Boston: Springer, [2] Horst R, Tuy H. Global Optimization, 3-rd Edition, New York: Springer, [3] Konno H. A Bilinear Programming: Part II. Applications of Bilinear Programming. Technical Report 71-10, Operations Research House, Stanford University, Stanford, CA. 5
6 [4] Konno H. A Cutting Plane Algorithm for Solving Bilinear Programs. Mathematical Programming 1976;11: [5] Nahapetyan A, Pardalos P. A Bilinear Relaxation Based Algorithm for Concave Piecewise Linear Network Flow Problems. Journal of Industrial and Management Optimization 3(1), pp , [6] Nahapetyan A, Pardalos P. Adaptive Dynamic Cost Updating Procedure for Solving Fixed Charge Network Flow Problems. Computational Optimization and Applications, submitted for publication. 6
Bilinear Modeling Solution Approach for Fixed Charged Network Flow Problems (Draft)
Bilinear Modeling Solution Approach for Fixed Charged Network Flow Problems (Draft) Steffen Rebennack 1, Artyom Nahapetyan 2, Panos M. Pardalos 1 1 Department of Industrial & Systems Engineering, Center
More informationAn algorithm for bilinear fractional problems
Laboratoire d Arithmétique, de Calcul formel et d Optimisation ESA - CNRS 6090 An algorithm for bilinear fractional problems Christian Malivert Rapport de recherche n 1998-03 Université de Limoges, 123
More informationAdvanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material
More informationIntroduction to Mathematical Programming IE406. Lecture 4. Dr. Ted Ralphs
Introduction to Mathematical Programming IE406 Lecture 4 Dr. Ted Ralphs IE406 Lecture 4 1 Reading for This Lecture Bertsimas 2.2-2.4 IE406 Lecture 4 2 The Two Crude Petroleum Example Revisited Recall the
More informationChapter 4 Concepts from Geometry
Chapter 4 Concepts from Geometry An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Line Segments The line segment between two points and in R n is the set of points on the straight line joining
More informationConvexity Theory and Gradient Methods
Convexity Theory and Gradient Methods Angelia Nedić angelia@illinois.edu ISE Department and Coordinated Science Laboratory University of Illinois at Urbana-Champaign Outline Convex Functions Optimality
More informationLecture Notes 2: The Simplex Algorithm
Algorithmic Methods 25/10/2010 Lecture Notes 2: The Simplex Algorithm Professor: Yossi Azar Scribe:Kiril Solovey 1 Introduction In this lecture we will present the Simplex algorithm, finish some unresolved
More informationDC Programming: A brief tutorial. Andres Munoz Medina Courant Institute of Mathematical Sciences
DC Programming: A brief tutorial. Andres Munoz Medina Courant Institute of Mathematical Sciences munoz@cims.nyu.edu Difference of Convex Functions Definition: A function exists f is said to be DC if there
More informationConvex Optimization Lecture 2
Convex Optimization Lecture 2 Today: Convex Analysis Center-of-mass Algorithm 1 Convex Analysis Convex Sets Definition: A set C R n is convex if for all x, y C and all 0 λ 1, λx + (1 λ)y C Operations that
More informationIntroduction to optimization
Introduction to optimization G. Ferrari Trecate Dipartimento di Ingegneria Industriale e dell Informazione Università degli Studi di Pavia Industrial Automation Ferrari Trecate (DIS) Optimization Industrial
More informationA PARAMETRIC SIMPLEX METHOD FOR OPTIMIZING A LINEAR FUNCTION OVER THE EFFICIENT SET OF A BICRITERIA LINEAR PROBLEM. 1.
ACTA MATHEMATICA VIETNAMICA Volume 21, Number 1, 1996, pp. 59 67 59 A PARAMETRIC SIMPLEX METHOD FOR OPTIMIZING A LINEAR FUNCTION OVER THE EFFICIENT SET OF A BICRITERIA LINEAR PROBLEM NGUYEN DINH DAN AND
More informationGlobal Minimization via Piecewise-Linear Underestimation
Journal of Global Optimization,, 1 9 (2004) c 2004 Kluwer Academic Publishers, Boston. Manufactured in The Netherlands. Global Minimization via Piecewise-Linear Underestimation O. L. MANGASARIAN olvi@cs.wisc.edu
More informationAn FPTAS for minimizing the product of two non-negative linear cost functions
Math. Program., Ser. A DOI 10.1007/s10107-009-0287-4 SHORT COMMUNICATION An FPTAS for minimizing the product of two non-negative linear cost functions Vineet Goyal Latife Genc-Kaya R. Ravi Received: 2
More informationProgramming, numerics and optimization
Programming, numerics and optimization Lecture C-4: Constrained optimization Łukasz Jankowski ljank@ippt.pan.pl Institute of Fundamental Technological Research Room 4.32, Phone +22.8261281 ext. 428 June
More information11 Linear Programming
11 Linear Programming 11.1 Definition and Importance The final topic in this course is Linear Programming. We say that a problem is an instance of linear programming when it can be effectively expressed
More informationConvexization in Markov Chain Monte Carlo
in Markov Chain Monte Carlo 1 IBM T. J. Watson Yorktown Heights, NY 2 Department of Aerospace Engineering Technion, Israel August 23, 2011 Problem Statement MCMC processes in general are governed by non
More information4 LINEAR PROGRAMMING (LP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1
4 LINEAR PROGRAMMING (LP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1 Mathematical programming (optimization) problem: min f (x) s.t. x X R n set of feasible solutions with linear objective function
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 2. Convex Optimization
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 2 Convex Optimization Shiqian Ma, MAT-258A: Numerical Optimization 2 2.1. Convex Optimization General optimization problem: min f 0 (x) s.t., f i
More informationGraphs that have the feasible bases of a given linear
Algorithmic Operations Research Vol.1 (2006) 46 51 Simplex Adjacency Graphs in Linear Optimization Gerard Sierksma and Gert A. Tijssen University of Groningen, Faculty of Economics, P.O. Box 800, 9700
More informationcontrol polytope. These points are manipulated by a descent method to compute a candidate global minimizer. The second method is described in Section
Some Heuristics and Test Problems for Nonconvex Quadratic Programming over a Simplex Ivo Nowak September 3, 1998 Keywords:global optimization, nonconvex quadratic programming, heuristics, Bezier methods,
More informationSolutions for Operations Research Final Exam
Solutions for Operations Research Final Exam. (a) The buffer stock is B = i a i = a + a + a + a + a + a 6 + a 7 = + + + + + + =. And the transportation tableau corresponding to the transshipment problem
More informationOptimization Methods. Final Examination. 1. There are 5 problems each w i t h 20 p o i n ts for a maximum of 100 points.
5.93 Optimization Methods Final Examination Instructions:. There are 5 problems each w i t h 2 p o i n ts for a maximum of points. 2. You are allowed to use class notes, your homeworks, solutions to homework
More informationIntroduction to Modern Control Systems
Introduction to Modern Control Systems Convex Optimization, Duality and Linear Matrix Inequalities Kostas Margellos University of Oxford AIMS CDT 2016-17 Introduction to Modern Control Systems November
More informationInteger Programming Theory
Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x
More information5.3 Cutting plane methods and Gomory fractional cuts
5.3 Cutting plane methods and Gomory fractional cuts (ILP) min c T x s.t. Ax b x 0integer feasible region X Assumption: a ij, c j and b i integer. Observation: The feasible region of an ILP can be described
More information6 Randomized rounding of semidefinite programs
6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can
More informationLinear Programming in Small Dimensions
Linear Programming in Small Dimensions Lekcija 7 sergio.cabello@fmf.uni-lj.si FMF Univerza v Ljubljani Edited from slides by Antoine Vigneron Outline linear programming, motivation and definition one dimensional
More informationConvex sets and convex functions
Convex sets and convex functions Convex optimization problems Convex sets and their examples Separating and supporting hyperplanes Projections on convex sets Convex functions, conjugate functions ECE 602,
More informationLinear Optimization. Andongwisye John. November 17, Linkoping University. Andongwisye John (Linkoping University) November 17, / 25
Linear Optimization Andongwisye John Linkoping University November 17, 2016 Andongwisye John (Linkoping University) November 17, 2016 1 / 25 Overview 1 Egdes, One-Dimensional Faces, Adjacency of Extreme
More informationSurrogate Gradient Algorithm for Lagrangian Relaxation 1,2
Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.
More informationReal life Problem. Review
Linear Programming The Modelling Cycle in Decision Maths Accept solution Real life Problem Yes No Review Make simplifying assumptions Compare the solution with reality is it realistic? Interpret the solution
More informationA Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems
A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems Keely L. Croxton Fisher College of Business The Ohio State University Bernard Gendron Département
More informationLECTURE NOTES Non-Linear Programming
CEE 6110 David Rosenberg p. 1 Learning Objectives LECTURE NOTES Non-Linear Programming 1. Write out the non-linear model formulation 2. Describe the difficulties of solving a non-linear programming model
More informationLecture 19 Subgradient Methods. November 5, 2008
Subgradient Methods November 5, 2008 Outline Lecture 19 Subgradients and Level Sets Subgradient Method Convergence and Convergence Rate Convex Optimization 1 Subgradients and Level Sets A vector s is a
More informationApplied Lagrange Duality for Constrained Optimization
Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity
More informationSYSTEMS OF NONLINEAR EQUATIONS
SYSTEMS OF NONLINEAR EQUATIONS Widely used in the mathematical modeling of real world phenomena. We introduce some numerical methods for their solution. For better intuition, we examine systems of two
More informationMathematical Programming and Research Methods (Part II)
Mathematical Programming and Research Methods (Part II) 4. Convexity and Optimization Massimiliano Pontil (based on previous lecture by Andreas Argyriou) 1 Today s Plan Convex sets and functions Types
More informationAdvanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 2 Review Dr. Ted Ralphs IE316 Quiz 2 Review 1 Reading for The Quiz Material covered in detail in lecture Bertsimas 4.1-4.5, 4.8, 5.1-5.5, 6.1-6.3 Material
More informationApplied Integer Programming
Applied Integer Programming D.S. Chen; R.G. Batson; Y. Dang Fahimeh 8.2 8.7 April 21, 2015 Context 8.2. Convex sets 8.3. Describing a bounded polyhedron 8.4. Describing unbounded polyhedron 8.5. Faces,
More informationResearch Interests Optimization:
Mitchell: Research interests 1 Research Interests Optimization: looking for the best solution from among a number of candidates. Prototypical optimization problem: min f(x) subject to g(x) 0 x X IR n Here,
More informationLinear and Integer Programming :Algorithms in the Real World. Related Optimization Problems. How important is optimization?
Linear and Integer Programming 15-853:Algorithms in the Real World Linear and Integer Programming I Introduction Geometric Interpretation Simplex Method Linear or Integer programming maximize z = c T x
More informationConic Duality. yyye
Conic Linear Optimization and Appl. MS&E314 Lecture Note #02 1 Conic Duality Yinyu Ye Department of Management Science and Engineering Stanford University Stanford, CA 94305, U.S.A. http://www.stanford.edu/
More informationInteger Programming ISE 418. Lecture 1. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 1 Dr. Ted Ralphs ISE 418 Lecture 1 1 Reading for This Lecture N&W Sections I.1.1-I.1.4 Wolsey Chapter 1 CCZ Chapter 2 ISE 418 Lecture 1 2 Mathematical Optimization Problems
More informationMath 5593 Linear Programming Lecture Notes
Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................
More informationPOLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if
POLYHEDRAL GEOMETRY Mathematical Programming Niels Lauritzen 7.9.2007 Convex functions and sets Recall that a subset C R n is convex if {λx + (1 λ)y 0 λ 1} C for every x, y C and 0 λ 1. A function f :
More informationLecture 2 - Introduction to Polytopes
Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.
More informationOn the Global Solution of Linear Programs with Linear Complementarity Constraints
On the Global Solution of Linear Programs with Linear Complementarity Constraints J. E. Mitchell 1 J. Hu 1 J.-S. Pang 2 K. P. Bennett 1 G. Kunapuli 1 1 Department of Mathematical Sciences RPI, Troy, NY
More informationNewton Polygons of L-Functions
Newton Polygons of L-Functions Phong Le Department of Mathematics University of California, Irvine June 2009/Ph.D. Defense Phong Le Newton Polygons of L-Functions 1/33 Laurent Polynomials Let q = p a where
More informationMATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS
MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS GRADO EN A.D.E. GRADO EN ECONOMÍA GRADO EN F.Y.C. ACADEMIC YEAR 2011-12 INDEX UNIT 1.- AN INTRODUCCTION TO OPTIMIZATION 2 UNIT 2.- NONLINEAR PROGRAMMING
More informationPrimal and Dual Methods for Optimisation over the Non-dominated Set of a Multi-objective Programme and Computing the Nadir Point
Primal and Dual Methods for Optimisation over the Non-dominated Set of a Multi-objective Programme and Computing the Nadir Point Ethan Liu Supervisor: Professor Matthias Ehrgott Lancaster University Outline
More informationChapter 15 Introduction to Linear Programming
Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of
More informationWeek 5. Convex Optimization
Week 5. Convex Optimization Lecturer: Prof. Santosh Vempala Scribe: Xin Wang, Zihao Li Feb. 9 and, 206 Week 5. Convex Optimization. The convex optimization formulation A general optimization problem is
More informationLinear programming and duality theory
Linear programming and duality theory Complements of Operations Research Giovanni Righini Linear Programming (LP) A linear program is defined by linear constraints, a linear objective function. Its variables
More informationCollege of Computer & Information Science Fall 2007 Northeastern University 14 September 2007
College of Computer & Information Science Fall 2007 Northeastern University 14 September 2007 CS G399: Algorithmic Power Tools I Scribe: Eric Robinson Lecture Outline: Linear Programming: Vertex Definitions
More information1. Lecture notes on bipartite matching February 4th,
1. Lecture notes on bipartite matching February 4th, 2015 6 1.1.1 Hall s Theorem Hall s theorem gives a necessary and sufficient condition for a bipartite graph to have a matching which saturates (or matches)
More informationProperties of Quadratic functions
Name Today s Learning Goals: #1 How do we determine the axis of symmetry and vertex of a quadratic function? Properties of Quadratic functions Date 5-1 Properties of a Quadratic Function A quadratic equation
More informationCS 435, 2018 Lecture 2, Date: 1 March 2018 Instructor: Nisheeth Vishnoi. Convex Programming and Efficiency
CS 435, 2018 Lecture 2, Date: 1 March 2018 Instructor: Nisheeth Vishnoi Convex Programming and Efficiency In this lecture, we formalize convex programming problem, discuss what it means to solve it efficiently
More informationFundamentals of Integer Programming
Fundamentals of Integer Programming Di Yuan Department of Information Technology, Uppsala University January 2018 Outline Definition of integer programming Formulating some classical problems with integer
More information9.5 Equivalence Relations
9.5 Equivalence Relations You know from your early study of fractions that each fraction has many equivalent forms. For example, 2, 2 4, 3 6, 2, 3 6, 5 30,... are all different ways to represent the same
More informationOptimality certificates for convex minimization and Helly numbers
Optimality certificates for convex minimization and Helly numbers Amitabh Basu Michele Conforti Gérard Cornuéjols Robert Weismantel Stefan Weltge October 20, 2016 Abstract We consider the problem of minimizing
More informationB553 Lecture 12: Global Optimization
B553 Lecture 12: Global Optimization Kris Hauser February 20, 2012 Most of the techniques we have examined in prior lectures only deal with local optimization, so that we can only guarantee convergence
More information1 Linear Programming. 1.1 Optimizion problems and convex polytopes 1 LINEAR PROGRAMMING
1 LINEAR PROGRAMMING 1 Linear Programming Now, we will talk a little bit about Linear Programming. We say that a problem is an instance of linear programming when it can be effectively expressed in the
More informationConvex Optimization MLSS 2015
Convex Optimization MLSS 2015 Constantine Caramanis The University of Texas at Austin The Optimization Problem minimize : f (x) subject to : x X. The Optimization Problem minimize : f (x) subject to :
More informationCopyright 2007 Pearson Addison-Wesley. All rights reserved. A. Levitin Introduction to the Design & Analysis of Algorithms, 2 nd ed., Ch.
Iterative Improvement Algorithm design technique for solving optimization problems Start with a feasible solution Repeat the following step until no improvement can be found: change the current feasible
More informationORIE 6300 Mathematical Programming I November 13, Lecture 23. max b T y. x 0 s 0. s.t. A T y + s = c
ORIE 63 Mathematical Programming I November 13, 214 Lecturer: David P. Williamson Lecture 23 Scribe: Mukadder Sevi Baltaoglu 1 Interior Point Methods Consider the standard primal and dual linear programs:
More informationConvex sets and convex functions
Convex sets and convex functions Convex optimization problems Convex sets and their examples Separating and supporting hyperplanes Projections on convex sets Convex functions, conjugate functions ECE 602,
More informationTMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions
Chalmers/GU Mathematics EXAM TMA946/MAN280 APPLIED OPTIMIZATION Date: 03 05 28 Time: House V, morning Aids: Text memory-less calculator Number of questions: 7; passed on one question requires 2 points
More informationAs a consequence of the operation, there are new incidences between edges and triangles that did not exist in K; see Figure II.9.
II.4 Surface Simplification 37 II.4 Surface Simplification In applications it is often necessary to simplify the data or its representation. One reason is measurement noise, which we would like to eliminate,
More informationCharacterizing Improving Directions Unconstrained Optimization
Final Review IE417 In the Beginning... In the beginning, Weierstrass's theorem said that a continuous function achieves a minimum on a compact set. Using this, we showed that for a convex set S and y not
More informationOptimized Choice of Parameters in interior-point methods for linear programming. Luiz-Rafael dos Santos
Optimized Choice of Parameters in interior-point methods for linear programming Luiz-Rafael dos Santos Joint work with F. Villas-Bôas, A. Oliveira and C. Perin 1st Brazilian Workshop on Interior Point
More informationUnconstrained Optimization Principles of Unconstrained Optimization Search Methods
1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization
More information15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs
15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest
More informationLecture 9. Semidefinite programming is linear programming where variables are entries in a positive semidefinite matrix.
CSE525: Randomized Algorithms and Probabilistic Analysis Lecture 9 Lecturer: Anna Karlin Scribe: Sonya Alexandrova and Keith Jia 1 Introduction to semidefinite programming Semidefinite programming is linear
More informationLecture 2 September 3
EE 381V: Large Scale Optimization Fall 2012 Lecture 2 September 3 Lecturer: Caramanis & Sanghavi Scribe: Hongbo Si, Qiaoyang Ye 2.1 Overview of the last Lecture The focus of the last lecture was to give
More informationInverse and Implicit functions
CHAPTER 3 Inverse and Implicit functions. Inverse Functions and Coordinate Changes Let U R d be a domain. Theorem. (Inverse function theorem). If ϕ : U R d is differentiable at a and Dϕ a is invertible,
More informationDavid G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer
David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms
More informationLecture 5: Properties of convex sets
Lecture 5: Properties of convex sets Rajat Mittal IIT Kanpur This week we will see properties of convex sets. These properties make convex sets special and are the reason why convex optimization problems
More information1 Linear programming relaxation
Cornell University, Fall 2010 CS 6820: Algorithms Lecture notes: Primal-dual min-cost bipartite matching August 27 30 1 Linear programming relaxation Recall that in the bipartite minimum-cost perfect matching
More informationMVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg
MVE165/MMG630, Integer linear programming algorithms Ann-Brith Strömberg 2009 04 15 Methods for ILP: Overview (Ch. 14.1) Enumeration Implicit enumeration: Branch and bound Relaxations Decomposition methods:
More informationChapter 3 Numerical Methods
Chapter 3 Numerical Methods Part 1 3.1 Linearization and Optimization of Functions of Vectors 1 Problem Notation 2 Outline 3.1.1 Linearization 3.1.2 Optimization of Objective Functions 3.1.3 Constrained
More informationDiscrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity
Discrete Optimization 2010 Lecture 5 Min-Cost Flows & Total Unimodularity Marc Uetz University of Twente m.uetz@utwente.nl Lecture 5: sheet 1 / 26 Marc Uetz Discrete Optimization Outline 1 Min-Cost Flows
More informationLinear Bilevel Programming With Upper Level Constraints Depending on the Lower Level Solution
Linear Bilevel Programming With Upper Level Constraints Depending on the Lower Level Solution Ayalew Getachew Mersha and Stephan Dempe October 17, 2005 Abstract Focus in the paper is on the definition
More informationA NEW SEQUENTIAL CUTTING PLANE ALGORITHM FOR SOLVING MIXED INTEGER NONLINEAR PROGRAMMING PROBLEMS
EVOLUTIONARY METHODS FOR DESIGN, OPTIMIZATION AND CONTROL P. Neittaanmäki, J. Périaux and T. Tuovinen (Eds.) c CIMNE, Barcelona, Spain 2007 A NEW SEQUENTIAL CUTTING PLANE ALGORITHM FOR SOLVING MIXED INTEGER
More informationCS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm. Instructor: Shaddin Dughmi
CS675: Convex and Combinatorial Optimization Spring 2018 Consequences of the Ellipsoid Algorithm Instructor: Shaddin Dughmi Outline 1 Recapping the Ellipsoid Method 2 Complexity of Convex Optimization
More information6. Lecture notes on matroid intersection
Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans May 2, 2017 6. Lecture notes on matroid intersection One nice feature about matroids is that a simple greedy algorithm
More informationLP-Modelling. dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven. January 30, 2008
LP-Modelling dr.ir. C.A.J. Hurkens Technische Universiteit Eindhoven January 30, 2008 1 Linear and Integer Programming After a brief check with the backgrounds of the participants it seems that the following
More informationMathematical and Algorithmic Foundations Linear Programming and Matchings
Adavnced Algorithms Lectures Mathematical and Algorithmic Foundations Linear Programming and Matchings Paul G. Spirakis Department of Computer Science University of Patras and Liverpool Paul G. Spirakis
More informationIntroduction to Optimization
Introduction to Optimization Second Order Optimization Methods Marc Toussaint U Stuttgart Planned Outline Gradient-based optimization (1st order methods) plain grad., steepest descent, conjugate grad.,
More information2. Optimization problems 6
6 2.1 Examples... 7... 8 2.3 Convex sets and functions... 9 2.4 Convex optimization problems... 10 2.1 Examples 7-1 An (NP-) optimization problem P 0 is defined as follows Each instance I P 0 has a feasibility
More informationThe Simplex Algorithm for LP, and an Open Problem
The Simplex Algorithm for LP, and an Open Problem Linear Programming: General Formulation Inputs: real-valued m x n matrix A, and vectors c in R n and b in R m Output: n-dimensional vector x There is one
More informationLinear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming
Linear Programming 3 describes a broad class of optimization tasks in which both the optimization criterion and the constraints are linear functions. Linear Programming consists of three parts: A set of
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationCS 372: Computational Geometry Lecture 10 Linear Programming in Fixed Dimension
CS 372: Computational Geometry Lecture 10 Linear Programming in Fixed Dimension Antoine Vigneron King Abdullah University of Science and Technology November 7, 2012 Antoine Vigneron (KAUST) CS 372 Lecture
More informationThe Branch-and-Sandwich Algorithm for Mixed-Integer Nonlinear Bilevel Problems
The Branch-and-Sandwich Algorithm for Mixed-Integer Nonlinear Bilevel Problems Polyxeni-M. Kleniati and Claire S. Adjiman MINLP Workshop June 2, 204, CMU Funding Bodies EPSRC & LEVERHULME TRUST OUTLINE
More information12.1 Formulation of General Perfect Matching
CSC5160: Combinatorial Optimization and Approximation Algorithms Topic: Perfect Matching Polytope Date: 22/02/2008 Lecturer: Lap Chi Lau Scribe: Yuk Hei Chan, Ling Ding and Xiaobing Wu In this lecture,
More informationAn Introduction to Bilevel Programming
An Introduction to Bilevel Programming Chris Fricke Department of Mathematics and Statistics University of Melbourne Outline What is Bilevel Programming? Origins of Bilevel Programming. Some Properties
More informationThe simplex method and the diameter of a 0-1 polytope
The simplex method and the diameter of a 0-1 polytope Tomonari Kitahara and Shinji Mizuno May 2012 Abstract We will derive two main results related to the primal simplex method for an LP on a 0-1 polytope.
More informationAlgorithms for finding the minimum cycle mean in the weighted directed graph
Computer Science Journal of Moldova, vol.6, no.1(16), 1998 Algorithms for finding the minimum cycle mean in the weighted directed graph D. Lozovanu C. Petic Abstract In this paper we study the problem
More informationLecture 19: Convex Non-Smooth Optimization. April 2, 2007
: Convex Non-Smooth Optimization April 2, 2007 Outline Lecture 19 Convex non-smooth problems Examples Subgradients and subdifferentials Subgradient properties Operations with subgradients and subdifferentials
More informationThe Simplex Algorithm
The Simplex Algorithm Uri Feige November 2011 1 The simplex algorithm The simplex algorithm was designed by Danzig in 1947. This write-up presents the main ideas involved. It is a slight update (mostly
More information