ONLY AVAILABLE IN ELECTRONIC FORM
|
|
- Ashlynn Phelps
- 5 years ago
- Views:
Transcription
1 MANAGEMENT SCIENCE doi /mnsc ec pp. ec1 ec7 e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 2008 INFORMS Electronic Companion Customized Bundle Pricing for Information Goods: A Nonlinear Mixed-Integer Programming Approach by Shin-yi Wu, Lorin M. Hitt, Pei-yu Chen, and G. Anandalingam, Management Science, doi /mnsc
2 Customized Bundle Pricing for Information Goods: A Nonlinear Mixed Integer Programming Approach On-line Supplement 1. Solution Approach 1.1. Lagrangian Relaxation By using the Lagrangian relaxation method, we can transform the primal problem (IP) mentioned above into the following Lagrangian relaxation problem (LR) where constraint (5) is dualized: Problem LR φ( a)= ( P B ) X MY + a (1 X ) (11) s.t. Max P, Si, Xi, Y ( ) i = 1,..., J i i i= 1,..., I = 1,..., J i = 1,..., J i= 1,..., I i = 1,..., J i S R - P Y, i = 1,..., I; = 1,..., J (12) i i S = ( R - P) X, i = 1,..., I (13) ( R - P) X 0, i = 1,..., I; = 1,..., J (14) i i X Y, i = 1,..., I; = 1,..., J (15) i S 0, i = 1,..., I (16) i P 0, = 1,..., J (17) X = 0 or 1, i = 1,..., I; = 1,..., J (18) i Y = 0 or 1, = 1,..., J (19) Since we relax only one constraint here, we have one type of corresponding Lagrangian Multiplier a i which must be nonnegative because the constraint we relaxed is an inequality constraint. Note that we can rearrange the terms in obective function of Problem LR and obtain the following obective function: Max φ a P 1,..., 1,..., B ai Xi MY J i I a = = i= 1,..., I i P, Si, Xi, Y ( )= [ ( ) ]+ (20) 1
3 This nonlinear integer programming problem LR is complex and hard to solve. However, it is amenable to simplification. First, given constraint (13) and (14), constraint (16) is redundant, so we can remove it. Second, constraint (13) simply defines the consumer surplus. If we replace S i in constraint (12) with k=1,,j (R ik -P k )X ik, we can remove constraint (13). After making these substitutions, we get an equivalent problem as follows: Problem LR ( a)= [ ( P B a ) X MY ]+ a (21) φ 1,..., 1,..., i i = J i= I i= 1,..., I i P, Xi, Y s.t. k= 1,..., J Max ( ) ( R - P ) X R - P Y, i = 1,..., I; = 1,..., J (22) ik k ik i ( R - P) X 0, i = 1,..., I; = 1,..., J i i (23) X Y, i = 1,..., I; = 1,..., J (24) i P 0, = 1,..., J (25) X = 0 or 1, i = 1,..., I; = 1,..., J (26) i Y = 0 or 1, = 1,..., J (27) Note LR can be decomposed into J different problems except for the complicated constraint (22). Unfortunately (22) cannot be dualized because the resulting Lagrangian obective function will have the nonlinear terms of P X i, P k X ik and P Y as well as the double summation over J, which makes it impossible to decompose the problem further into J different subproblems. As a result, we decide to drop constraint (22) and solve a relaxed version of LR, which we call Problem R. Problem R Va ( )= [ ( P B a ) X MY ]+ a (28) s.t. Max P, Xi, Y = 1,..., J i= 1,..., I i i i= 1,..., I i ( R - P) X 0, i = 1,..., I; = 1,..., J (29) i i X Y, i = 1,..., I; = 1,..., J i (30) P 0, = 1,..., J (31) X = 0 or 1, i = 1,..., I; = 1,..., J (32) i Y = 0 or 1, = 1,..., J (33) 2
4 It is not hard to see that Problem R is actually the summation of J independent subproblems (one for each, as shown in (34)-(39)) plus a constant term i=1,,i a i. Subproblem Max P, Xi, Y s.t. i= 1,..., I ( P B a ) X MY (34) i i ( R - P) X 0, i = 1,..., I (35) i i X Y, i = 1,..., I (36) i P 0, (37) X = 0 or 1, i = 1,..., I (38) Y i = 0 or 1, (39) Each such subproblem is actually equivalent to the following setting. Given each customer s reservation price for the bundle of goods (i.e., R i ), we are trying to maximize vendor s total profit by deciding whether to offer the bundle of goods (i.e., determine Y ) and the price of the bundle P. The fixed cost of offering the bundle is M while the marginal bundle cost to provide this bundle to customer i is B + a i (a i is Lagrangian Multiplier). Again, each customer is assumed to maximize her surplus and she is willing to purchase the bundle (set X i to 1) when the surplus is nonnegative (constraint (35)) Algorithmic Intuition Under this Subproblem, the vendor should offer a bundle of goods if the maximum revenue i=1,,i (P - B - a i )X i achievable by providing this bundle is larger than the fixed cost M. In other words, a single variable Y appears in each Subproblem. If it is 0, then all X i are 0 and the value of the obective function is 0. Therefore we will choose to set Y to 1 only if improvement is possible (the best obective function value for Y =1 is larger than 0). In addition, the vendor is willing to sell the bundle with goods to customer i only when it is profitable: P - B - a i 0, which would make it optimal for X i to be 1. Keeping these in mind, we first try to find the most profitable price P and use it to calculate the maximum revenue i=1,,i (P - B - a i )X i for each subproblem. If the resulting revenue is larger than M, the vendor should offer the bundle by setting Y = 1. We then set X i of those profitable and willing-to-buy 3
5 customers to 1, others to 0, and the optimal obective function value of this subproblem is equal to (maximum revenue - M). On the other hand, if the maximum revenue achievable is smaller than M, the vendor should not offer the bundle. We then simply set all decision variables P and Y to 0, which in turn set all X i to zero, and the optimal obective function value of this subproblem is equal to 0. To derive the most profitable price P, let us first sort the reservation price R i of all potential customers in descending order and call them R 1, R 2,, R I. Obviously, if we set the price above R 1, no customer will be willing to buy and the maximum revenue achievable is 0. So setting the price above R 1 is not profitable. Now assume we set the price between R k and R k+1 (R k > R k+1 ), then at most k customers will be willing to buy. We soon discover that if we raise the price a bit to R k, the total number of willingto-buy customers will not change while the maximum profit achievable will rise for sure. In other words, any price setting between R k and R k+1 is no better than R k, and when considering the best price, we only have to take R 1, R 2,, R I as possible candidates. To this point, we successfully decreased the set of possible optimal pricing P from a continuous set to a discrete and finite one. Since the set is finite (with size I), we can simply choose the price that maximizes revenue from this set. After the above analysis, we can design the following algorithm to solve Problem R, the further relaxed problem of Problem LR where constraint (22) is removed. Step 1. =1,,J { choose the best price P among R 1, R 2,, R I by comparing the maximum revenue achievable; if maximum revenue M, set Y = 1 and set X i of those profitable and willing-to-buy customers to 1, others to 0; else set P, X i, and Y to 0 } Step 2. Add the obective function value of J subproblems together plus a constant term i=1,,i a i. This is the obective function value of Problem R. 4
6 1.3. The Dual Problem and the Subgradient Method According to the algorithm proposed above, although we could not directly solve the Lagrangian relaxation problem Problem LR, we can successfully solve Problem R. According to the weak Lagrangian duality theorem, the optimal obective value of Problem LR is an upper bound of the optimal obective value of Primal Problem IP. And since Problem R is a relaxation of Problem LR, the optimal obective value of Problem R is also an upper bound of the optimal obective value of IP. We then construct the following dual problem to calculate the tightest upper bound and solve the dual problem by using the subgradient method (more details about the subgradient method can be found in the Bertsekas, 1999): min V(a) (D) s.t. a 0. where V(a) is Problem R. Let the vector S be a subgradient of V(a) at (a 1, a 2,, a I ). In iteration k of the subgradient optimization procedure, the multiplier a is updated by a k+1 = a k - α k S k where S k = 1 - =1,,J X i. The step size α k k is determined by α k h V( a ) φ = β where φ h is the primal obective value, which k 2 S we get from a heuristic solution (a lower bound of the optimal primal obective value), and β is a scalar, 0 β 2. We run the subgradient method for 1000 iterations, the value of β is reset every 200 iterations, from 2 to 1.6, 1.2, 0.8 and finally 0.4, and it is divided by 2 after 40 iterations without upper bound improvement. The initial value of the multiplier vector a is chosen to be the 0 vector and φ h is always updated to the best lower bound obtained at the time. After the implementation of the subgradient optimization procedure, while we didn t observe a point with zero subgradient, we have upper bound on the optimal obective value of the primal problem. However, as expected, no primal feasible solution was found in the process due to the structure and complexity of this bundle pricing problem since the subdivision into J independent subproblems means 5
7 that constraint (5) is likely to be violated. In the following subsection, we try to utilize the Lagrangian solutions to develop a heuristic algorithm to get the primal feasible solutions Getting Primal Feasible Solutions For each iteration of the subgradient method, we get a Lagrangian solution which is affected by the updating of the Lagrangian Multiplier a. But because we relaxed constraints (5) and (22) above, the Lagrangian solutions we get for Problem R may not be feasible for the original primal problem. However, from past experiences, we know that the degree of violations is often low and it is possible to obtain a good primal feasible solution using heuristics which start from a Lagrangian solution. We now describe this heuristic in which we use P and Y from the Lagrangian solutions as given and recalculate X i (and also S i ) to get the primal feasible solutions. We can use this to generate a sequence of primal feasible solutions by adusting the sequence of Lagrangian feasible solutions we get from the subgradient method. Algorithm Bundling : Step 1. For each potential customer i, choose the bundle k (only those with Y = 1) with the largest positive surplus. Set X ik to 1 and other X i to 0. (This enforces constraint (22) and constraint (5)). Step 2. For each offered bundle (those with Y = 1), calculate its revenue achieved i=1,,i (P - B )X i. Choose the bundle k with the smallest revenue. If the revenue is smaller than M, we do not offer this bundle by setting the corresponding P, X i, and Y to 0 and go to step 1. (The rationale behind this is that based on our primal obective function (1), it is reasonable for us to offer the bundle only when the revenue from the bundle could cover the fixed cost M.) Step 3. Calculate consumer surplus S i according to constraint (3). More information about Lagrangian relaxation and subgradient methods can be found in Fisher (1981), which was recently chosen as one of the top ten most influential Management Science paper in the past 50 years, and Nowak (2005). 6
8 References: Bertsekas, D. Nonlinear Programming. Nashua, NH: Athena Scientific, pp , Fisher, M. The LaGrangian-Relaxation Method for Solving Integer Programming-Problems, Management Science (27:1), 1981, pp Nowak, I. Lagrangian Decomposition of Block-separable Mixed-Integer All-Quadratic programs, Mathematical Programming 102, 2005, pp
Lecture 4 Duality and Decomposition Techniques
Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian
More informationAdvanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs
Advanced Operations Research Techniques IE316 Quiz 2 Review Dr. Ted Ralphs IE316 Quiz 2 Review 1 Reading for The Quiz Material covered in detail in lecture Bertsimas 4.1-4.5, 4.8, 5.1-5.5, 6.1-6.3 Material
More informationLagrangian Relaxation: An overview
Discrete Math for Bioinformatics WS 11/12:, by A. Bockmayr/K. Reinert, 22. Januar 2013, 13:27 4001 Lagrangian Relaxation: An overview Sources for this lecture: D. Bertsimas and J. Tsitsiklis: Introduction
More informationSurrogate Gradient Algorithm for Lagrangian Relaxation 1,2
Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.
More informationImproving Dual Bound for Stochastic MILP Models Using Sensitivity Analysis
Improving Dual Bound for Stochastic MILP Models Using Sensitivity Analysis Vijay Gupta Ignacio E. Grossmann Department of Chemical Engineering Carnegie Mellon University, Pittsburgh Bora Tarhan ExxonMobil
More information56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998
56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998 Part A: Answer any four of the five problems. (15 points each) 1. Transportation problem 2. Integer LP Model Formulation
More information15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs
15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest
More informationGate Sizing by Lagrangian Relaxation Revisited
Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007
More informationNonlinear Programming
Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,
More informationBCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D.
Linear Programming Module Outline Introduction The Linear Programming Model Examples of Linear Programming Problems Developing Linear Programming Models Graphical Solution to LP Problems The Simplex Method
More informationLagrangean relaxation - exercises
Lagrangean relaxation - exercises Giovanni Righini Set covering We start from the following Set Covering Problem instance: min z = x + 2x 2 + x + 2x 4 + x 5 x + x 2 + x 4 x 2 + x x 2 + x 4 + x 5 x + x
More informationPart 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm
In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.
More informationApplied Lagrange Duality for Constrained Optimization
Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity
More informationAn Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation
Dhaka Univ. J. Sci. 61(2): 135-140, 2013 (July) An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation M. Babul Hasan and Md. Toha De epartment of Mathematics, Dhaka
More informationMVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg
MVE165/MMG630, Integer linear programming algorithms Ann-Brith Strömberg 2009 04 15 Methods for ILP: Overview (Ch. 14.1) Enumeration Implicit enumeration: Branch and bound Relaxations Decomposition methods:
More informationNetwork Flows. 7. Multicommodity Flows Problems. Fall 2010 Instructor: Dr. Masoud Yaghini
In the name of God Network Flows 7. Multicommodity Flows Problems 7.2 Lagrangian Relaxation Approach Fall 2010 Instructor: Dr. Masoud Yaghini The multicommodity flow problem formulation: We associate nonnegative
More informationInteger Programming: Algorithms - 2
Week 8 Integer Programming: Algorithms - 2 OPR 992 Applied Mathematical Programming OPR 992 - Applied Mathematical Programming - p. 1/13 Introduction Valid Inequalities Chvatal-Gomory Procedure Gomory
More information5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY
5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY 5.1 DUALITY Associated with every linear programming problem (the primal) is another linear programming problem called its dual. If the primal involves
More informationConvex Optimization and Machine Learning
Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction
More informationSimulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016
Simulation Lecture O Optimization: Linear Programming Saeed Bastani April 06 Outline of the course Linear Programming ( lecture) Integer Programming ( lecture) Heuristics and Metaheursitics (3 lectures)
More informationLagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound
Lagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound Francisco Trespalacios, Ignacio E. Grossmann Department of Chemical
More informationLink Dimensioning and LSP Optimization for MPLS Networks Supporting DiffServ EF and BE Classes
Link Dimensioning and LSP Optimization for MPLS Networks Supporting DiffServ EF and BE Classes Kehang Wu Douglas S. Reeves Capacity Planning for QoS DiffServ + MPLS QoS in core networks DiffServ provides
More informationSelected Topics in Column Generation
Selected Topics in Column Generation February 1, 2007 Choosing a solver for the Master Solve in the dual space(kelly s method) by applying a cutting plane algorithm In the bundle method(lemarechal), a
More informationMVE165/MMG631 Linear and integer optimization with applications Lecture 9 Discrete optimization: theory and algorithms
MVE165/MMG631 Linear and integer optimization with applications Lecture 9 Discrete optimization: theory and algorithms Ann-Brith Strömberg 2018 04 24 Lecture 9 Linear and integer optimization with applications
More informationHomework 2: Multi-unit combinatorial auctions (due Nov. 7 before class)
CPS 590.1 - Linear and integer programming Homework 2: Multi-unit combinatorial auctions (due Nov. 7 before class) Please read the rules for assignments on the course web page. Contact Vince (conitzer@cs.duke.edu)
More informationSparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual
Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal
More informationIntroduction to Linear Programming. Algorithmic and Geometric Foundations of Optimization
Introduction to Linear Programming Algorithmic and Geometric Foundations of Optimization Optimization and Linear Programming Mathematical programming is a class of methods for solving problems which ask
More informationLECTURE 18 LECTURE OUTLINE
LECTURE 18 LECTURE OUTLINE Generalized polyhedral approximation methods Combined cutting plane and simplicial decomposition methods Lecture based on the paper D. P. Bertsekas and H. Yu, A Unifying Polyhedral
More informationThe Size Robust Multiple Knapsack Problem
MASTER THESIS ICA-3251535 The Size Robust Multiple Knapsack Problem Branch and Price for the Separate and Combined Recovery Decomposition Model Author: D.D. Tönissen, Supervisors: dr. ir. J.M. van den
More informationCOMS 4771 Support Vector Machines. Nakul Verma
COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron
More informationInteger Programming. Xi Chen. Department of Management Science and Engineering International Business School Beijing Foreign Studies University
Integer Programming Xi Chen Department of Management Science and Engineering International Business School Beijing Foreign Studies University Xi Chen (chenxi0109@bfsu.edu.cn) Integer Programming 1 / 42
More informationDISCRETE CONVEX ANALYSIS
DISCRETE CONVEX ANALYSIS o KAZUO MUROTA University of Tokyo; PRESTO, JST Tokyo, Japan Society for Industrial and Applied Mathematics Philadelphia List of Figures Notation Preface xi xiii xxi 1 Introduction
More informationMATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS
MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS GRADO EN A.D.E. GRADO EN ECONOMÍA GRADO EN F.Y.C. ACADEMIC YEAR 2011-12 INDEX UNIT 1.- AN INTRODUCCTION TO OPTIMIZATION 2 UNIT 2.- NONLINEAR PROGRAMMING
More informationLagrangian Relaxation
Lagrangian Relaxation Ş. İlker Birbil March 6, 2016 where Our general optimization problem in this lecture is in the maximization form: maximize f(x) subject to x F, F = {x R n : g i (x) 0, i = 1,...,
More information3 INTEGER LINEAR PROGRAMMING
3 INTEGER LINEAR PROGRAMMING PROBLEM DEFINITION Integer linear programming problem (ILP) of the decision variables x 1,..,x n : (ILP) subject to minimize c x j j n j= 1 a ij x j x j 0 x j integer n j=
More informationChapter II. Linear Programming
1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION
More informationStochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition
Stochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition Xiang Li Process Systems Engineering Laboratory Department of Chemical Engineering Massachusetts
More informationModified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with
Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Parallelization in Non-hierarchical Analytical Target Cascading Yongsu Jung Department of Mechanical Engineering,
More informationFinite Math Linear Programming 1 May / 7
Linear Programming Finite Math 1 May 2017 Finite Math Linear Programming 1 May 2017 1 / 7 General Description of Linear Programming Finite Math Linear Programming 1 May 2017 2 / 7 General Description of
More informationParallel and Distributed Graph Cuts by Dual Decomposition
Parallel and Distributed Graph Cuts by Dual Decomposition Presented by Varun Gulshan 06 July, 2010 What is Dual Decomposition? What is Dual Decomposition? It is a technique for optimization: usually used
More informationAlgorithms for Integer Programming
Algorithms for Integer Programming Laura Galli November 9, 2016 Unlike linear programming problems, integer programming problems are very difficult to solve. In fact, no efficient general algorithm is
More informationIntroduction. Linear because it requires linear functions. Programming as synonymous of planning.
LINEAR PROGRAMMING Introduction Development of linear programming was among the most important scientific advances of mid-20th cent. Most common type of applications: allocate limited resources to competing
More informationCONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year
CONLIN & MMA solvers Pierre DUYSINX LTAS Automotive Engineering Academic year 2018-2019 1 CONLIN METHOD 2 LAY-OUT CONLIN SUBPROBLEMS DUAL METHOD APPROACH FOR CONLIN SUBPROBLEMS SEQUENTIAL QUADRATIC PROGRAMMING
More informationTreatment Planning Optimization for VMAT, Tomotherapy and Cyberknife
Treatment Planning Optimization for VMAT, Tomotherapy and Cyberknife Kerem Akartunalı Department of Management Science Strathclyde Business School Joint work with: Vicky Mak-Hau and Thu Tran 14 July 2015
More informationDavid G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer
David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms
More informationChapter 15 Introduction to Linear Programming
Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of
More informationConstrained optimization
Constrained optimization A general constrained optimization problem has the form where The Lagrangian function is given by Primal and dual optimization problems Primal: Dual: Weak duality: Strong duality:
More informationA Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems
A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems Keely L. Croxton Fisher College of Business The Ohio State University Bernard Gendron Département
More information11 Linear Programming
11 Linear Programming 11.1 Definition and Importance The final topic in this course is Linear Programming. We say that a problem is an instance of linear programming when it can be effectively expressed
More informationOutline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014
5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38
More information1 Linear programming relaxation
Cornell University, Fall 2010 CS 6820: Algorithms Lecture notes: Primal-dual min-cost bipartite matching August 27 30 1 Linear programming relaxation Recall that in the bipartite minimum-cost perfect matching
More informationTMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions
Chalmers/GU Mathematics EXAM TMA946/MAN280 APPLIED OPTIMIZATION Date: 03 05 28 Time: House V, morning Aids: Text memory-less calculator Number of questions: 7; passed on one question requires 2 points
More informationPRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction
PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem
More informationDepartment of Electrical and Computer Engineering
LAGRANGIAN RELAXATION FOR GATE IMPLEMENTATION SELECTION Yi-Le Huang, Jiang Hu and Weiping Shi Department of Electrical and Computer Engineering Texas A&M University OUTLINE Introduction and motivation
More informationApproximation Algorithms
Approximation Algorithms Group Members: 1. Geng Xue (A0095628R) 2. Cai Jingli (A0095623B) 3. Xing Zhe (A0095644W) 4. Zhu Xiaolu (A0109657W) 5. Wang Zixiao (A0095670X) 6. Jiao Qing (A0095637R) 7. Zhang
More informationDETERMINISTIC OPERATIONS RESEARCH
DETERMINISTIC OPERATIONS RESEARCH Models and Methods in Optimization Linear DAVID J. RADER, JR. Rose-Hulman Institute of Technology Department of Mathematics Terre Haute, IN WILEY A JOHN WILEY & SONS,
More informationRead: H&L chapters 1-6
Viterbi School of Engineering Daniel J. Epstein Department of Industrial and Systems Engineering ISE 330: Introduction to Operations Research Fall 2006 (Oct 16): Midterm Review http://www-scf.usc.edu/~ise330
More informationChapter 7. Linear Programming Models: Graphical and Computer Methods
Chapter 7 Linear Programming Models: Graphical and Computer Methods To accompany Quantitative Analysis for Management, Eleventh Edition, by Render, Stair, and Hanna Power Point slides created by Brian
More information5.3 Cutting plane methods and Gomory fractional cuts
5.3 Cutting plane methods and Gomory fractional cuts (ILP) min c T x s.t. Ax b x 0integer feasible region X Assumption: a ij, c j and b i integer. Observation: The feasible region of an ILP can be described
More informationCOLUMN GENERATION IN LINEAR PROGRAMMING
COLUMN GENERATION IN LINEAR PROGRAMMING EXAMPLE: THE CUTTING STOCK PROBLEM A certain material (e.g. lumber) is stocked in lengths of 9, 4, and 6 feet, with respective costs of $5, $9, and $. An order for
More informationApproximation Algorithms
Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A 4 credit unit course Part of Theoretical Computer Science courses at the Laboratory of Mathematics There will be 4 hours
More informationDual-fitting analysis of Greedy for Set Cover
Dual-fitting analysis of Greedy for Set Cover We showed earlier that the greedy algorithm for set cover gives a H n approximation We will show that greedy produces a solution of cost at most H n OPT LP
More informationBenders in a nutshell Matteo Fischetti, University of Padova
Benders in a nutshell Matteo Fischetti, University of Padova ODS 2017, Sorrento, September 2017 1 Benders decomposition The original Benders decomposition from the 1960s uses two distinct ingredients for
More informationSupport Vector Machines. James McInerney Adapted from slides by Nakul Verma
Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake
More informationto the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics
An Application of Lagrangian Relaxation to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics M lardalen University SE-721
More informationCivil Engineering Systems Analysis Lecture XIV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics
Civil Engineering Systems Analysis Lecture XIV Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics Today s Learning Objectives Dual 2 Linear Programming Dual Problem 3
More informationLECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach
LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction
More informationInteger Programming ISE 418. Lecture 7. Dr. Ted Ralphs
Integer Programming ISE 418 Lecture 7 Dr. Ted Ralphs ISE 418 Lecture 7 1 Reading for This Lecture Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Wolsey Chapter 7 CCZ Chapter 1 Constraint
More informationx ji = s i, i N, (1.1)
Dual Ascent Methods. DUAL ASCENT In this chapter we focus on the minimum cost flow problem minimize subject to (i,j) A {j (i,j) A} a ij x ij x ij {j (j,i) A} (MCF) x ji = s i, i N, (.) b ij x ij c ij,
More informationLecture 7: Support Vector Machine
Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each
More informationDecomposition and Dynamic Cut Generation in Integer Linear Programming
Decomposition and Dynamic Cut Generation in Integer Linear Programming M.V. Galati Advanced Analytics - Operations R & D, SAS Institute, Chesterbrook, PA 908 Ted K. Ralphs Department of Industrial and
More informationApproximation Techniques for Utilitarian Mechanism Design
Approximation Techniques for Utilitarian Mechanism Design Department of Computer Science RWTH Aachen Germany joint work with Patrick Briest and Piotr Krysta 05/16/2006 1 Introduction to Utilitarian Mechanism
More informationIE521 Convex Optimization Introduction
IE521 Convex Optimization Introduction Instructor: Niao He Jan 18, 2017 1 About Me Assistant Professor, UIUC, 2016 Ph.D. in Operations Research, M.S. in Computational Sci. & Eng. Georgia Tech, 2010 2015
More informationGENERALIZED ASSIGNMENT PROBLEM
Chapter 3 GENERALIZED ASSIGNMENT PROBLEM 1 1 A part of this chapter appears in International Journal of Computational and Mathematical Sciences Vol.3:8, 418-423, 2009 as A new heuristic approach for the
More informationLinear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming
Linear Programming 3 describes a broad class of optimization tasks in which both the optimization criterion and the constraints are linear functions. Linear Programming consists of three parts: A set of
More informationLaGO - A solver for mixed integer nonlinear programming
LaGO - A solver for mixed integer nonlinear programming Ivo Nowak June 1 2005 Problem formulation MINLP: min f(x, y) s.t. g(x, y) 0 h(x, y) = 0 x [x, x] y [y, y] integer MINLP: - n
More informationArtificial Intelligence
Artificial Intelligence Combinatorial Optimization G. Guérard Department of Nouvelles Energies Ecole Supérieur d Ingénieurs Léonard de Vinci Lecture 1 GG A.I. 1/34 Outline 1 Motivation 2 Geometric resolution
More informationBenders Decomposition
Benders Decomposition Using projections to solve problems thst@man.dtu.dk DTU-Management Technical University of Denmark 1 Outline Introduction Using projections Benders decomposition Simple plant location
More informationPerceptron Learning Algorithm
Perceptron Learning Algorithm An iterative learning algorithm that can find linear threshold function to partition linearly separable set of points. Assume zero threshold value. 1) w(0) = arbitrary, j=1,
More informationSome Advanced Topics in Linear Programming
Some Advanced Topics in Linear Programming Matthew J. Saltzman July 2, 995 Connections with Algebra and Geometry In this section, we will explore how some of the ideas in linear programming, duality theory,
More informationNetwork Disruption Recovery for Multiple Pairs of Shortest Paths
Network Disruption ecovery for Multiple Pairs of Shortest Paths Lindong Liu Department of IELM The Hong Kong University of Science and Technology Email: ldliu@ust.hk Xiangtong Qi Department of IELM The
More informationMath 5593 Linear Programming Final Exam
Math 5593 Linear Programming Final Exam Department of Mathematical and Statistical Sciences University of Colorado Denver, Fall 2013 Name: Points: /30 This exam consists of 6 problems, and each problem
More informationSVMs for Structured Output. Andrea Vedaldi
SVMs for Structured Output Andrea Vedaldi SVM struct Tsochantaridis Hofmann Joachims Altun 04 Extending SVMs 3 Extending SVMs SVM = parametric function arbitrary input binary output 3 Extending SVMs SVM
More informationAssignment 3b: The traveling salesman problem
Chalmers University of Technology MVE165 University of Gothenburg MMG631 Mathematical Sciences Linear and integer optimization Optimization with applications Emil Gustavsson Assignment information Ann-Brith
More informationLECTURE 5: DUAL PROBLEMS AND KERNELS. * Most of the slides in this lecture are from
LECTURE 5: DUAL PROBLEMS AND KERNELS * Most of the slides in this lecture are from http://www.robots.ox.ac.uk/~az/lectures/ml Optimization Loss function Loss functions SVM review PRIMAL-DUAL PROBLEM Max-min
More informationAn Introduction to Dual Ascent Heuristics
An Introduction to Dual Ascent Heuristics Introduction A substantial proportion of Combinatorial Optimisation Problems (COPs) are essentially pure or mixed integer linear programming. COPs are in general
More informationIntroduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs
Introduction to Mathematical Programming IE496 Final Review Dr. Ted Ralphs IE496 Final Review 1 Course Wrap-up: Chapter 2 In the introduction, we discussed the general framework of mathematical modeling
More information56:272 Integer Programming & Network Flows Final Exam -- December 16, 1997
56:272 Integer Programming & Network Flows Final Exam -- December 16, 1997 Answer #1 and any five of the remaining six problems! possible score 1. Multiple Choice 25 2. Traveling Salesman Problem 15 3.
More informationSolutions for Operations Research Final Exam
Solutions for Operations Research Final Exam. (a) The buffer stock is B = i a i = a + a + a + a + a + a 6 + a 7 = + + + + + + =. And the transportation tableau corresponding to the transshipment problem
More informationME 555: Distributed Optimization
ME 555: Distributed Optimization Duke University Spring 2015 1 Administrative Course: ME 555: Distributed Optimization (Spring 2015) Instructor: Time: Location: Office hours: Website: Soomin Lee (email:
More informationInteger Programming Theory
Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x
More informationA Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems
A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems E. O. Effanga Department of Mathematics/Statistics and Comp. Science University of Calabar P.M.B. 1115, Calabar, Cross
More informationUnconstrained Optimization Principles of Unconstrained Optimization Search Methods
1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization
More informationA Generic Separation Algorithm and Its Application to the Vehicle Routing Problem
A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem Presented by: Ted Ralphs Joint work with: Leo Kopman Les Trotter Bill Pulleyblank 1 Outline of Talk Introduction Description
More informationApproximation Algorithms
Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours
More informationDecomposition in Integer Linear Programming
Decomposition in Integer Linear Programming T.K. Ralphs M.V. Galati December 3, 00 Abstract Both cutting plane methods and traditional decomposition methods are procedures that compute a bound on the optimal
More informationPrimal Dual Schema Approach to the Labeling Problem with Applications to TSP
1 Primal Dual Schema Approach to the Labeling Problem with Applications to TSP Colin Brown, Simon Fraser University Instructor: Ramesh Krishnamurti The Metric Labeling Problem has many applications, especially
More informationLinear Programming Problems
Linear Programming Problems Two common formulations of linear programming (LP) problems are: min Subject to: 1,,, 1,2,,;, max Subject to: 1,,, 1,2,,;, Linear Programming Problems The standard LP problem
More informationIntroduction to Constrained Optimization
Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR
More informationDecomposition in Integer Linear Programming
Decomposition in Integer Linear Programming M.V. Galati Advanced Analytics - Operations R & D, SAS Institute, Chesterbrook, PA 9 Ted K. Ralphs Department of Industrial and System Engineering, Lehigh University,
More information