ONLY AVAILABLE IN ELECTRONIC FORM

Size: px
Start display at page:

Download "ONLY AVAILABLE IN ELECTRONIC FORM"

Transcription

1 MANAGEMENT SCIENCE doi /mnsc ec pp. ec1 ec7 e-companion ONLY AVAILABLE IN ELECTRONIC FORM informs 2008 INFORMS Electronic Companion Customized Bundle Pricing for Information Goods: A Nonlinear Mixed-Integer Programming Approach by Shin-yi Wu, Lorin M. Hitt, Pei-yu Chen, and G. Anandalingam, Management Science, doi /mnsc

2 Customized Bundle Pricing for Information Goods: A Nonlinear Mixed Integer Programming Approach On-line Supplement 1. Solution Approach 1.1. Lagrangian Relaxation By using the Lagrangian relaxation method, we can transform the primal problem (IP) mentioned above into the following Lagrangian relaxation problem (LR) where constraint (5) is dualized: Problem LR φ( a)= ( P B ) X MY + a (1 X ) (11) s.t. Max P, Si, Xi, Y ( ) i = 1,..., J i i i= 1,..., I = 1,..., J i = 1,..., J i= 1,..., I i = 1,..., J i S R - P Y, i = 1,..., I; = 1,..., J (12) i i S = ( R - P) X, i = 1,..., I (13) ( R - P) X 0, i = 1,..., I; = 1,..., J (14) i i X Y, i = 1,..., I; = 1,..., J (15) i S 0, i = 1,..., I (16) i P 0, = 1,..., J (17) X = 0 or 1, i = 1,..., I; = 1,..., J (18) i Y = 0 or 1, = 1,..., J (19) Since we relax only one constraint here, we have one type of corresponding Lagrangian Multiplier a i which must be nonnegative because the constraint we relaxed is an inequality constraint. Note that we can rearrange the terms in obective function of Problem LR and obtain the following obective function: Max φ a P 1,..., 1,..., B ai Xi MY J i I a = = i= 1,..., I i P, Si, Xi, Y ( )= [ ( ) ]+ (20) 1

3 This nonlinear integer programming problem LR is complex and hard to solve. However, it is amenable to simplification. First, given constraint (13) and (14), constraint (16) is redundant, so we can remove it. Second, constraint (13) simply defines the consumer surplus. If we replace S i in constraint (12) with k=1,,j (R ik -P k )X ik, we can remove constraint (13). After making these substitutions, we get an equivalent problem as follows: Problem LR ( a)= [ ( P B a ) X MY ]+ a (21) φ 1,..., 1,..., i i = J i= I i= 1,..., I i P, Xi, Y s.t. k= 1,..., J Max ( ) ( R - P ) X R - P Y, i = 1,..., I; = 1,..., J (22) ik k ik i ( R - P) X 0, i = 1,..., I; = 1,..., J i i (23) X Y, i = 1,..., I; = 1,..., J (24) i P 0, = 1,..., J (25) X = 0 or 1, i = 1,..., I; = 1,..., J (26) i Y = 0 or 1, = 1,..., J (27) Note LR can be decomposed into J different problems except for the complicated constraint (22). Unfortunately (22) cannot be dualized because the resulting Lagrangian obective function will have the nonlinear terms of P X i, P k X ik and P Y as well as the double summation over J, which makes it impossible to decompose the problem further into J different subproblems. As a result, we decide to drop constraint (22) and solve a relaxed version of LR, which we call Problem R. Problem R Va ( )= [ ( P B a ) X MY ]+ a (28) s.t. Max P, Xi, Y = 1,..., J i= 1,..., I i i i= 1,..., I i ( R - P) X 0, i = 1,..., I; = 1,..., J (29) i i X Y, i = 1,..., I; = 1,..., J i (30) P 0, = 1,..., J (31) X = 0 or 1, i = 1,..., I; = 1,..., J (32) i Y = 0 or 1, = 1,..., J (33) 2

4 It is not hard to see that Problem R is actually the summation of J independent subproblems (one for each, as shown in (34)-(39)) plus a constant term i=1,,i a i. Subproblem Max P, Xi, Y s.t. i= 1,..., I ( P B a ) X MY (34) i i ( R - P) X 0, i = 1,..., I (35) i i X Y, i = 1,..., I (36) i P 0, (37) X = 0 or 1, i = 1,..., I (38) Y i = 0 or 1, (39) Each such subproblem is actually equivalent to the following setting. Given each customer s reservation price for the bundle of goods (i.e., R i ), we are trying to maximize vendor s total profit by deciding whether to offer the bundle of goods (i.e., determine Y ) and the price of the bundle P. The fixed cost of offering the bundle is M while the marginal bundle cost to provide this bundle to customer i is B + a i (a i is Lagrangian Multiplier). Again, each customer is assumed to maximize her surplus and she is willing to purchase the bundle (set X i to 1) when the surplus is nonnegative (constraint (35)) Algorithmic Intuition Under this Subproblem, the vendor should offer a bundle of goods if the maximum revenue i=1,,i (P - B - a i )X i achievable by providing this bundle is larger than the fixed cost M. In other words, a single variable Y appears in each Subproblem. If it is 0, then all X i are 0 and the value of the obective function is 0. Therefore we will choose to set Y to 1 only if improvement is possible (the best obective function value for Y =1 is larger than 0). In addition, the vendor is willing to sell the bundle with goods to customer i only when it is profitable: P - B - a i 0, which would make it optimal for X i to be 1. Keeping these in mind, we first try to find the most profitable price P and use it to calculate the maximum revenue i=1,,i (P - B - a i )X i for each subproblem. If the resulting revenue is larger than M, the vendor should offer the bundle by setting Y = 1. We then set X i of those profitable and willing-to-buy 3

5 customers to 1, others to 0, and the optimal obective function value of this subproblem is equal to (maximum revenue - M). On the other hand, if the maximum revenue achievable is smaller than M, the vendor should not offer the bundle. We then simply set all decision variables P and Y to 0, which in turn set all X i to zero, and the optimal obective function value of this subproblem is equal to 0. To derive the most profitable price P, let us first sort the reservation price R i of all potential customers in descending order and call them R 1, R 2,, R I. Obviously, if we set the price above R 1, no customer will be willing to buy and the maximum revenue achievable is 0. So setting the price above R 1 is not profitable. Now assume we set the price between R k and R k+1 (R k > R k+1 ), then at most k customers will be willing to buy. We soon discover that if we raise the price a bit to R k, the total number of willingto-buy customers will not change while the maximum profit achievable will rise for sure. In other words, any price setting between R k and R k+1 is no better than R k, and when considering the best price, we only have to take R 1, R 2,, R I as possible candidates. To this point, we successfully decreased the set of possible optimal pricing P from a continuous set to a discrete and finite one. Since the set is finite (with size I), we can simply choose the price that maximizes revenue from this set. After the above analysis, we can design the following algorithm to solve Problem R, the further relaxed problem of Problem LR where constraint (22) is removed. Step 1. =1,,J { choose the best price P among R 1, R 2,, R I by comparing the maximum revenue achievable; if maximum revenue M, set Y = 1 and set X i of those profitable and willing-to-buy customers to 1, others to 0; else set P, X i, and Y to 0 } Step 2. Add the obective function value of J subproblems together plus a constant term i=1,,i a i. This is the obective function value of Problem R. 4

6 1.3. The Dual Problem and the Subgradient Method According to the algorithm proposed above, although we could not directly solve the Lagrangian relaxation problem Problem LR, we can successfully solve Problem R. According to the weak Lagrangian duality theorem, the optimal obective value of Problem LR is an upper bound of the optimal obective value of Primal Problem IP. And since Problem R is a relaxation of Problem LR, the optimal obective value of Problem R is also an upper bound of the optimal obective value of IP. We then construct the following dual problem to calculate the tightest upper bound and solve the dual problem by using the subgradient method (more details about the subgradient method can be found in the Bertsekas, 1999): min V(a) (D) s.t. a 0. where V(a) is Problem R. Let the vector S be a subgradient of V(a) at (a 1, a 2,, a I ). In iteration k of the subgradient optimization procedure, the multiplier a is updated by a k+1 = a k - α k S k where S k = 1 - =1,,J X i. The step size α k k is determined by α k h V( a ) φ = β where φ h is the primal obective value, which k 2 S we get from a heuristic solution (a lower bound of the optimal primal obective value), and β is a scalar, 0 β 2. We run the subgradient method for 1000 iterations, the value of β is reset every 200 iterations, from 2 to 1.6, 1.2, 0.8 and finally 0.4, and it is divided by 2 after 40 iterations without upper bound improvement. The initial value of the multiplier vector a is chosen to be the 0 vector and φ h is always updated to the best lower bound obtained at the time. After the implementation of the subgradient optimization procedure, while we didn t observe a point with zero subgradient, we have upper bound on the optimal obective value of the primal problem. However, as expected, no primal feasible solution was found in the process due to the structure and complexity of this bundle pricing problem since the subdivision into J independent subproblems means 5

7 that constraint (5) is likely to be violated. In the following subsection, we try to utilize the Lagrangian solutions to develop a heuristic algorithm to get the primal feasible solutions Getting Primal Feasible Solutions For each iteration of the subgradient method, we get a Lagrangian solution which is affected by the updating of the Lagrangian Multiplier a. But because we relaxed constraints (5) and (22) above, the Lagrangian solutions we get for Problem R may not be feasible for the original primal problem. However, from past experiences, we know that the degree of violations is often low and it is possible to obtain a good primal feasible solution using heuristics which start from a Lagrangian solution. We now describe this heuristic in which we use P and Y from the Lagrangian solutions as given and recalculate X i (and also S i ) to get the primal feasible solutions. We can use this to generate a sequence of primal feasible solutions by adusting the sequence of Lagrangian feasible solutions we get from the subgradient method. Algorithm Bundling : Step 1. For each potential customer i, choose the bundle k (only those with Y = 1) with the largest positive surplus. Set X ik to 1 and other X i to 0. (This enforces constraint (22) and constraint (5)). Step 2. For each offered bundle (those with Y = 1), calculate its revenue achieved i=1,,i (P - B )X i. Choose the bundle k with the smallest revenue. If the revenue is smaller than M, we do not offer this bundle by setting the corresponding P, X i, and Y to 0 and go to step 1. (The rationale behind this is that based on our primal obective function (1), it is reasonable for us to offer the bundle only when the revenue from the bundle could cover the fixed cost M.) Step 3. Calculate consumer surplus S i according to constraint (3). More information about Lagrangian relaxation and subgradient methods can be found in Fisher (1981), which was recently chosen as one of the top ten most influential Management Science paper in the past 50 years, and Nowak (2005). 6

8 References: Bertsekas, D. Nonlinear Programming. Nashua, NH: Athena Scientific, pp , Fisher, M. The LaGrangian-Relaxation Method for Solving Integer Programming-Problems, Management Science (27:1), 1981, pp Nowak, I. Lagrangian Decomposition of Block-separable Mixed-Integer All-Quadratic programs, Mathematical Programming 102, 2005, pp

Lecture 4 Duality and Decomposition Techniques

Lecture 4 Duality and Decomposition Techniques Lecture 4 Duality and Decomposition Techniques Jie Lu (jielu@kth.se) Richard Combes Alexandre Proutiere Automatic Control, KTH September 19, 2013 Consider the primal problem Lagrange Duality Lagrangian

More information

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs

Advanced Operations Research Techniques IE316. Quiz 2 Review. Dr. Ted Ralphs Advanced Operations Research Techniques IE316 Quiz 2 Review Dr. Ted Ralphs IE316 Quiz 2 Review 1 Reading for The Quiz Material covered in detail in lecture Bertsimas 4.1-4.5, 4.8, 5.1-5.5, 6.1-6.3 Material

More information

Lagrangian Relaxation: An overview

Lagrangian Relaxation: An overview Discrete Math for Bioinformatics WS 11/12:, by A. Bockmayr/K. Reinert, 22. Januar 2013, 13:27 4001 Lagrangian Relaxation: An overview Sources for this lecture: D. Bertsimas and J. Tsitsiklis: Introduction

More information

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2

Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 Surrogate Gradient Algorithm for Lagrangian Relaxation 1,2 X. Zhao 3, P. B. Luh 4, and J. Wang 5 Communicated by W.B. Gong and D. D. Yao 1 This paper is dedicated to Professor Yu-Chi Ho for his 65th birthday.

More information

Improving Dual Bound for Stochastic MILP Models Using Sensitivity Analysis

Improving Dual Bound for Stochastic MILP Models Using Sensitivity Analysis Improving Dual Bound for Stochastic MILP Models Using Sensitivity Analysis Vijay Gupta Ignacio E. Grossmann Department of Chemical Engineering Carnegie Mellon University, Pittsburgh Bora Tarhan ExxonMobil

More information

56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998

56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998 56:272 Integer Programming & Network Flows Final Examination -- December 14, 1998 Part A: Answer any four of the five problems. (15 points each) 1. Transportation problem 2. Integer LP Model Formulation

More information

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs 15.082J and 6.855J Lagrangian Relaxation 2 Algorithms Application to LPs 1 The Constrained Shortest Path Problem (1,10) 2 (1,1) 4 (2,3) (1,7) 1 (10,3) (1,2) (10,1) (5,7) 3 (12,3) 5 (2,2) 6 Find the shortest

More information

Gate Sizing by Lagrangian Relaxation Revisited

Gate Sizing by Lagrangian Relaxation Revisited Gate Sizing by Lagrangian Relaxation Revisited Jia Wang, Debasish Das, and Hai Zhou Electrical Engineering and Computer Science Northwestern University Evanston, Illinois, United States October 17, 2007

More information

Nonlinear Programming

Nonlinear Programming Nonlinear Programming SECOND EDITION Dimitri P. Bertsekas Massachusetts Institute of Technology WWW site for book Information and Orders http://world.std.com/~athenasc/index.html Athena Scientific, Belmont,

More information

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D.

BCN Decision and Risk Analysis. Syed M. Ahmed, Ph.D. Linear Programming Module Outline Introduction The Linear Programming Model Examples of Linear Programming Problems Developing Linear Programming Models Graphical Solution to LP Problems The Simplex Method

More information

Lagrangean relaxation - exercises

Lagrangean relaxation - exercises Lagrangean relaxation - exercises Giovanni Righini Set covering We start from the following Set Covering Problem instance: min z = x + 2x 2 + x + 2x 4 + x 5 x + x 2 + x 4 x 2 + x x 2 + x 4 + x 5 x + x

More information

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm

Part 4. Decomposition Algorithms Dantzig-Wolf Decomposition Algorithm In the name of God Part 4. 4.1. Dantzig-Wolf Decomposition Algorithm Spring 2010 Instructor: Dr. Masoud Yaghini Introduction Introduction Real world linear programs having thousands of rows and columns.

More information

Applied Lagrange Duality for Constrained Optimization

Applied Lagrange Duality for Constrained Optimization Applied Lagrange Duality for Constrained Optimization Robert M. Freund February 10, 2004 c 2004 Massachusetts Institute of Technology. 1 1 Overview The Practical Importance of Duality Review of Convexity

More information

An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation

An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation Dhaka Univ. J. Sci. 61(2): 135-140, 2013 (July) An Improved Subgradiend Optimization Technique for Solving IPs with Lagrangean Relaxation M. Babul Hasan and Md. Toha De epartment of Mathematics, Dhaka

More information

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg

MVE165/MMG630, Applied Optimization Lecture 8 Integer linear programming algorithms. Ann-Brith Strömberg MVE165/MMG630, Integer linear programming algorithms Ann-Brith Strömberg 2009 04 15 Methods for ILP: Overview (Ch. 14.1) Enumeration Implicit enumeration: Branch and bound Relaxations Decomposition methods:

More information

Network Flows. 7. Multicommodity Flows Problems. Fall 2010 Instructor: Dr. Masoud Yaghini

Network Flows. 7. Multicommodity Flows Problems. Fall 2010 Instructor: Dr. Masoud Yaghini In the name of God Network Flows 7. Multicommodity Flows Problems 7.2 Lagrangian Relaxation Approach Fall 2010 Instructor: Dr. Masoud Yaghini The multicommodity flow problem formulation: We associate nonnegative

More information

Integer Programming: Algorithms - 2

Integer Programming: Algorithms - 2 Week 8 Integer Programming: Algorithms - 2 OPR 992 Applied Mathematical Programming OPR 992 - Applied Mathematical Programming - p. 1/13 Introduction Valid Inequalities Chvatal-Gomory Procedure Gomory

More information

5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY

5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY 5. DUAL LP, SOLUTION INTERPRETATION, AND POST-OPTIMALITY 5.1 DUALITY Associated with every linear programming problem (the primal) is another linear programming problem called its dual. If the primal involves

More information

Convex Optimization and Machine Learning

Convex Optimization and Machine Learning Convex Optimization and Machine Learning Mengliu Zhao Machine Learning Reading Group School of Computing Science Simon Fraser University March 12, 2014 Mengliu Zhao SFU-MLRG March 12, 2014 1 / 25 Introduction

More information

Simulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016

Simulation. Lecture O1 Optimization: Linear Programming. Saeed Bastani April 2016 Simulation Lecture O Optimization: Linear Programming Saeed Bastani April 06 Outline of the course Linear Programming ( lecture) Integer Programming ( lecture) Heuristics and Metaheursitics (3 lectures)

More information

Lagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound

Lagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound Lagrangean Relaxation of the Hull-Reformulation of Linear Generalized Disjunctive Programs and its use in Disjunctive Branch and Bound Francisco Trespalacios, Ignacio E. Grossmann Department of Chemical

More information

Link Dimensioning and LSP Optimization for MPLS Networks Supporting DiffServ EF and BE Classes

Link Dimensioning and LSP Optimization for MPLS Networks Supporting DiffServ EF and BE Classes Link Dimensioning and LSP Optimization for MPLS Networks Supporting DiffServ EF and BE Classes Kehang Wu Douglas S. Reeves Capacity Planning for QoS DiffServ + MPLS QoS in core networks DiffServ provides

More information

Selected Topics in Column Generation

Selected Topics in Column Generation Selected Topics in Column Generation February 1, 2007 Choosing a solver for the Master Solve in the dual space(kelly s method) by applying a cutting plane algorithm In the bundle method(lemarechal), a

More information

MVE165/MMG631 Linear and integer optimization with applications Lecture 9 Discrete optimization: theory and algorithms

MVE165/MMG631 Linear and integer optimization with applications Lecture 9 Discrete optimization: theory and algorithms MVE165/MMG631 Linear and integer optimization with applications Lecture 9 Discrete optimization: theory and algorithms Ann-Brith Strömberg 2018 04 24 Lecture 9 Linear and integer optimization with applications

More information

Homework 2: Multi-unit combinatorial auctions (due Nov. 7 before class)

Homework 2: Multi-unit combinatorial auctions (due Nov. 7 before class) CPS 590.1 - Linear and integer programming Homework 2: Multi-unit combinatorial auctions (due Nov. 7 before class) Please read the rules for assignments on the course web page. Contact Vince (conitzer@cs.duke.edu)

More information

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual

Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Sparse Optimization Lecture: Proximal Operator/Algorithm and Lagrange Dual Instructor: Wotao Yin July 2013 online discussions on piazza.com Those who complete this lecture will know learn the proximal

More information

Introduction to Linear Programming. Algorithmic and Geometric Foundations of Optimization

Introduction to Linear Programming. Algorithmic and Geometric Foundations of Optimization Introduction to Linear Programming Algorithmic and Geometric Foundations of Optimization Optimization and Linear Programming Mathematical programming is a class of methods for solving problems which ask

More information

LECTURE 18 LECTURE OUTLINE

LECTURE 18 LECTURE OUTLINE LECTURE 18 LECTURE OUTLINE Generalized polyhedral approximation methods Combined cutting plane and simplicial decomposition methods Lecture based on the paper D. P. Bertsekas and H. Yu, A Unifying Polyhedral

More information

The Size Robust Multiple Knapsack Problem

The Size Robust Multiple Knapsack Problem MASTER THESIS ICA-3251535 The Size Robust Multiple Knapsack Problem Branch and Price for the Separate and Combined Recovery Decomposition Model Author: D.D. Tönissen, Supervisors: dr. ir. J.M. van den

More information

COMS 4771 Support Vector Machines. Nakul Verma

COMS 4771 Support Vector Machines. Nakul Verma COMS 4771 Support Vector Machines Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake bound for the perceptron

More information

Integer Programming. Xi Chen. Department of Management Science and Engineering International Business School Beijing Foreign Studies University

Integer Programming. Xi Chen. Department of Management Science and Engineering International Business School Beijing Foreign Studies University Integer Programming Xi Chen Department of Management Science and Engineering International Business School Beijing Foreign Studies University Xi Chen (chenxi0109@bfsu.edu.cn) Integer Programming 1 / 42

More information

DISCRETE CONVEX ANALYSIS

DISCRETE CONVEX ANALYSIS DISCRETE CONVEX ANALYSIS o KAZUO MUROTA University of Tokyo; PRESTO, JST Tokyo, Japan Society for Industrial and Applied Mathematics Philadelphia List of Figures Notation Preface xi xiii xxi 1 Introduction

More information

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS

MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS MATHEMATICS II: COLLECTION OF EXERCISES AND PROBLEMS GRADO EN A.D.E. GRADO EN ECONOMÍA GRADO EN F.Y.C. ACADEMIC YEAR 2011-12 INDEX UNIT 1.- AN INTRODUCCTION TO OPTIMIZATION 2 UNIT 2.- NONLINEAR PROGRAMMING

More information

Lagrangian Relaxation

Lagrangian Relaxation Lagrangian Relaxation Ş. İlker Birbil March 6, 2016 where Our general optimization problem in this lecture is in the maximization form: maximize f(x) subject to x F, F = {x R n : g i (x) 0, i = 1,...,

More information

3 INTEGER LINEAR PROGRAMMING

3 INTEGER LINEAR PROGRAMMING 3 INTEGER LINEAR PROGRAMMING PROBLEM DEFINITION Integer linear programming problem (ILP) of the decision variables x 1,..,x n : (ILP) subject to minimize c x j j n j= 1 a ij x j x j 0 x j integer n j=

More information

Chapter II. Linear Programming

Chapter II. Linear Programming 1 Chapter II Linear Programming 1. Introduction 2. Simplex Method 3. Duality Theory 4. Optimality Conditions 5. Applications (QP & SLP) 6. Sensitivity Analysis 7. Interior Point Methods 1 INTRODUCTION

More information

Stochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition

Stochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition Stochastic Separable Mixed-Integer Nonlinear Programming via Nonconvex Generalized Benders Decomposition Xiang Li Process Systems Engineering Laboratory Department of Chemical Engineering Massachusetts

More information

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with

Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Modified Augmented Lagrangian Coordination and Alternating Direction Method of Multipliers with Parallelization in Non-hierarchical Analytical Target Cascading Yongsu Jung Department of Mechanical Engineering,

More information

Finite Math Linear Programming 1 May / 7

Finite Math Linear Programming 1 May / 7 Linear Programming Finite Math 1 May 2017 Finite Math Linear Programming 1 May 2017 1 / 7 General Description of Linear Programming Finite Math Linear Programming 1 May 2017 2 / 7 General Description of

More information

Parallel and Distributed Graph Cuts by Dual Decomposition

Parallel and Distributed Graph Cuts by Dual Decomposition Parallel and Distributed Graph Cuts by Dual Decomposition Presented by Varun Gulshan 06 July, 2010 What is Dual Decomposition? What is Dual Decomposition? It is a technique for optimization: usually used

More information

Algorithms for Integer Programming

Algorithms for Integer Programming Algorithms for Integer Programming Laura Galli November 9, 2016 Unlike linear programming problems, integer programming problems are very difficult to solve. In fact, no efficient general algorithm is

More information

Introduction. Linear because it requires linear functions. Programming as synonymous of planning.

Introduction. Linear because it requires linear functions. Programming as synonymous of planning. LINEAR PROGRAMMING Introduction Development of linear programming was among the most important scientific advances of mid-20th cent. Most common type of applications: allocate limited resources to competing

More information

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year

CONLIN & MMA solvers. Pierre DUYSINX LTAS Automotive Engineering Academic year CONLIN & MMA solvers Pierre DUYSINX LTAS Automotive Engineering Academic year 2018-2019 1 CONLIN METHOD 2 LAY-OUT CONLIN SUBPROBLEMS DUAL METHOD APPROACH FOR CONLIN SUBPROBLEMS SEQUENTIAL QUADRATIC PROGRAMMING

More information

Treatment Planning Optimization for VMAT, Tomotherapy and Cyberknife

Treatment Planning Optimization for VMAT, Tomotherapy and Cyberknife Treatment Planning Optimization for VMAT, Tomotherapy and Cyberknife Kerem Akartunalı Department of Management Science Strathclyde Business School Joint work with: Vicky Mak-Hau and Thu Tran 14 July 2015

More information

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer

David G. Luenberger Yinyu Ye. Linear and Nonlinear. Programming. Fourth Edition. ö Springer David G. Luenberger Yinyu Ye Linear and Nonlinear Programming Fourth Edition ö Springer Contents 1 Introduction 1 1.1 Optimization 1 1.2 Types of Problems 2 1.3 Size of Problems 5 1.4 Iterative Algorithms

More information

Chapter 15 Introduction to Linear Programming

Chapter 15 Introduction to Linear Programming Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2015 Wei-Ta Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of

More information

Constrained optimization

Constrained optimization Constrained optimization A general constrained optimization problem has the form where The Lagrangian function is given by Primal and dual optimization problems Primal: Dual: Weak duality: Strong duality:

More information

A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems

A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems A Comparison of Mixed-Integer Programming Models for Non-Convex Piecewise Linear Cost Minimization Problems Keely L. Croxton Fisher College of Business The Ohio State University Bernard Gendron Département

More information

11 Linear Programming

11 Linear Programming 11 Linear Programming 11.1 Definition and Importance The final topic in this course is Linear Programming. We say that a problem is an instance of linear programming when it can be effectively expressed

More information

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014 5/2/24 Outline CS38 Introduction to Algorithms Lecture 5 May 2, 24 Linear programming simplex algorithm LP duality ellipsoid algorithm * slides from Kevin Wayne May 2, 24 CS38 Lecture 5 May 2, 24 CS38

More information

1 Linear programming relaxation

1 Linear programming relaxation Cornell University, Fall 2010 CS 6820: Algorithms Lecture notes: Primal-dual min-cost bipartite matching August 27 30 1 Linear programming relaxation Recall that in the bipartite minimum-cost perfect matching

More information

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions

TMA946/MAN280 APPLIED OPTIMIZATION. Exam instructions Chalmers/GU Mathematics EXAM TMA946/MAN280 APPLIED OPTIMIZATION Date: 03 05 28 Time: House V, morning Aids: Text memory-less calculator Number of questions: 7; passed on one question requires 2 points

More information

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction

PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING. 1. Introduction PRIMAL-DUAL INTERIOR POINT METHOD FOR LINEAR PROGRAMMING KELLER VANDEBOGERT AND CHARLES LANNING 1. Introduction Interior point methods are, put simply, a technique of optimization where, given a problem

More information

Department of Electrical and Computer Engineering

Department of Electrical and Computer Engineering LAGRANGIAN RELAXATION FOR GATE IMPLEMENTATION SELECTION Yi-Le Huang, Jiang Hu and Weiping Shi Department of Electrical and Computer Engineering Texas A&M University OUTLINE Introduction and motivation

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Group Members: 1. Geng Xue (A0095628R) 2. Cai Jingli (A0095623B) 3. Xing Zhe (A0095644W) 4. Zhu Xiaolu (A0109657W) 5. Wang Zixiao (A0095670X) 6. Jiao Qing (A0095637R) 7. Zhang

More information

DETERMINISTIC OPERATIONS RESEARCH

DETERMINISTIC OPERATIONS RESEARCH DETERMINISTIC OPERATIONS RESEARCH Models and Methods in Optimization Linear DAVID J. RADER, JR. Rose-Hulman Institute of Technology Department of Mathematics Terre Haute, IN WILEY A JOHN WILEY & SONS,

More information

Read: H&L chapters 1-6

Read: H&L chapters 1-6 Viterbi School of Engineering Daniel J. Epstein Department of Industrial and Systems Engineering ISE 330: Introduction to Operations Research Fall 2006 (Oct 16): Midterm Review http://www-scf.usc.edu/~ise330

More information

Chapter 7. Linear Programming Models: Graphical and Computer Methods

Chapter 7. Linear Programming Models: Graphical and Computer Methods Chapter 7 Linear Programming Models: Graphical and Computer Methods To accompany Quantitative Analysis for Management, Eleventh Edition, by Render, Stair, and Hanna Power Point slides created by Brian

More information

5.3 Cutting plane methods and Gomory fractional cuts

5.3 Cutting plane methods and Gomory fractional cuts 5.3 Cutting plane methods and Gomory fractional cuts (ILP) min c T x s.t. Ax b x 0integer feasible region X Assumption: a ij, c j and b i integer. Observation: The feasible region of an ILP can be described

More information

COLUMN GENERATION IN LINEAR PROGRAMMING

COLUMN GENERATION IN LINEAR PROGRAMMING COLUMN GENERATION IN LINEAR PROGRAMMING EXAMPLE: THE CUTTING STOCK PROBLEM A certain material (e.g. lumber) is stocked in lengths of 9, 4, and 6 feet, with respective costs of $5, $9, and $. An order for

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A 4 credit unit course Part of Theoretical Computer Science courses at the Laboratory of Mathematics There will be 4 hours

More information

Dual-fitting analysis of Greedy for Set Cover

Dual-fitting analysis of Greedy for Set Cover Dual-fitting analysis of Greedy for Set Cover We showed earlier that the greedy algorithm for set cover gives a H n approximation We will show that greedy produces a solution of cost at most H n OPT LP

More information

Benders in a nutshell Matteo Fischetti, University of Padova

Benders in a nutshell Matteo Fischetti, University of Padova Benders in a nutshell Matteo Fischetti, University of Padova ODS 2017, Sorrento, September 2017 1 Benders decomposition The original Benders decomposition from the 1960s uses two distinct ingredients for

More information

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma

Support Vector Machines. James McInerney Adapted from slides by Nakul Verma Support Vector Machines James McInerney Adapted from slides by Nakul Verma Last time Decision boundaries for classification Linear decision boundary (linear classification) The Perceptron algorithm Mistake

More information

to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics

to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics An Application of Lagrangian Relaxation to the Traveling Salesman Problem 1 Susanne Timsj Applied Optimization and Modeling Group (TOM) Department of Mathematics and Physics M lardalen University SE-721

More information

Civil Engineering Systems Analysis Lecture XIV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics

Civil Engineering Systems Analysis Lecture XIV. Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics Civil Engineering Systems Analysis Lecture XIV Instructor: Prof. Naveen Eluru Department of Civil Engineering and Applied Mechanics Today s Learning Objectives Dual 2 Linear Programming Dual Problem 3

More information

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach

LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION. 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach LECTURE 13: SOLUTION METHODS FOR CONSTRAINED OPTIMIZATION 1. Primal approach 2. Penalty and barrier methods 3. Dual approach 4. Primal-dual approach Basic approaches I. Primal Approach - Feasible Direction

More information

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs

Integer Programming ISE 418. Lecture 7. Dr. Ted Ralphs Integer Programming ISE 418 Lecture 7 Dr. Ted Ralphs ISE 418 Lecture 7 1 Reading for This Lecture Nemhauser and Wolsey Sections II.3.1, II.3.6, II.4.1, II.4.2, II.5.4 Wolsey Chapter 7 CCZ Chapter 1 Constraint

More information

x ji = s i, i N, (1.1)

x ji = s i, i N, (1.1) Dual Ascent Methods. DUAL ASCENT In this chapter we focus on the minimum cost flow problem minimize subject to (i,j) A {j (i,j) A} a ij x ij x ij {j (j,i) A} (MCF) x ji = s i, i N, (.) b ij x ij c ij,

More information

Lecture 7: Support Vector Machine

Lecture 7: Support Vector Machine Lecture 7: Support Vector Machine Hien Van Nguyen University of Houston 9/28/2017 Separating hyperplane Red and green dots can be separated by a separating hyperplane Two classes are separable, i.e., each

More information

Decomposition and Dynamic Cut Generation in Integer Linear Programming

Decomposition and Dynamic Cut Generation in Integer Linear Programming Decomposition and Dynamic Cut Generation in Integer Linear Programming M.V. Galati Advanced Analytics - Operations R & D, SAS Institute, Chesterbrook, PA 908 Ted K. Ralphs Department of Industrial and

More information

Approximation Techniques for Utilitarian Mechanism Design

Approximation Techniques for Utilitarian Mechanism Design Approximation Techniques for Utilitarian Mechanism Design Department of Computer Science RWTH Aachen Germany joint work with Patrick Briest and Piotr Krysta 05/16/2006 1 Introduction to Utilitarian Mechanism

More information

IE521 Convex Optimization Introduction

IE521 Convex Optimization Introduction IE521 Convex Optimization Introduction Instructor: Niao He Jan 18, 2017 1 About Me Assistant Professor, UIUC, 2016 Ph.D. in Operations Research, M.S. in Computational Sci. & Eng. Georgia Tech, 2010 2015

More information

GENERALIZED ASSIGNMENT PROBLEM

GENERALIZED ASSIGNMENT PROBLEM Chapter 3 GENERALIZED ASSIGNMENT PROBLEM 1 1 A part of this chapter appears in International Journal of Computational and Mathematical Sciences Vol.3:8, 418-423, 2009 as A new heuristic approach for the

More information

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming

Linear Programming. Linear Programming. Linear Programming. Example: Profit Maximization (1/4) Iris Hui-Ru Jiang Fall Linear programming Linear Programming 3 describes a broad class of optimization tasks in which both the optimization criterion and the constraints are linear functions. Linear Programming consists of three parts: A set of

More information

LaGO - A solver for mixed integer nonlinear programming

LaGO - A solver for mixed integer nonlinear programming LaGO - A solver for mixed integer nonlinear programming Ivo Nowak June 1 2005 Problem formulation MINLP: min f(x, y) s.t. g(x, y) 0 h(x, y) = 0 x [x, x] y [y, y] integer MINLP: - n

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Combinatorial Optimization G. Guérard Department of Nouvelles Energies Ecole Supérieur d Ingénieurs Léonard de Vinci Lecture 1 GG A.I. 1/34 Outline 1 Motivation 2 Geometric resolution

More information

Benders Decomposition

Benders Decomposition Benders Decomposition Using projections to solve problems thst@man.dtu.dk DTU-Management Technical University of Denmark 1 Outline Introduction Using projections Benders decomposition Simple plant location

More information

Perceptron Learning Algorithm

Perceptron Learning Algorithm Perceptron Learning Algorithm An iterative learning algorithm that can find linear threshold function to partition linearly separable set of points. Assume zero threshold value. 1) w(0) = arbitrary, j=1,

More information

Some Advanced Topics in Linear Programming

Some Advanced Topics in Linear Programming Some Advanced Topics in Linear Programming Matthew J. Saltzman July 2, 995 Connections with Algebra and Geometry In this section, we will explore how some of the ideas in linear programming, duality theory,

More information

Network Disruption Recovery for Multiple Pairs of Shortest Paths

Network Disruption Recovery for Multiple Pairs of Shortest Paths Network Disruption ecovery for Multiple Pairs of Shortest Paths Lindong Liu Department of IELM The Hong Kong University of Science and Technology Email: ldliu@ust.hk Xiangtong Qi Department of IELM The

More information

Math 5593 Linear Programming Final Exam

Math 5593 Linear Programming Final Exam Math 5593 Linear Programming Final Exam Department of Mathematical and Statistical Sciences University of Colorado Denver, Fall 2013 Name: Points: /30 This exam consists of 6 problems, and each problem

More information

SVMs for Structured Output. Andrea Vedaldi

SVMs for Structured Output. Andrea Vedaldi SVMs for Structured Output Andrea Vedaldi SVM struct Tsochantaridis Hofmann Joachims Altun 04 Extending SVMs 3 Extending SVMs SVM = parametric function arbitrary input binary output 3 Extending SVMs SVM

More information

Assignment 3b: The traveling salesman problem

Assignment 3b: The traveling salesman problem Chalmers University of Technology MVE165 University of Gothenburg MMG631 Mathematical Sciences Linear and integer optimization Optimization with applications Emil Gustavsson Assignment information Ann-Brith

More information

LECTURE 5: DUAL PROBLEMS AND KERNELS. * Most of the slides in this lecture are from

LECTURE 5: DUAL PROBLEMS AND KERNELS. * Most of the slides in this lecture are from LECTURE 5: DUAL PROBLEMS AND KERNELS * Most of the slides in this lecture are from http://www.robots.ox.ac.uk/~az/lectures/ml Optimization Loss function Loss functions SVM review PRIMAL-DUAL PROBLEM Max-min

More information

An Introduction to Dual Ascent Heuristics

An Introduction to Dual Ascent Heuristics An Introduction to Dual Ascent Heuristics Introduction A substantial proportion of Combinatorial Optimisation Problems (COPs) are essentially pure or mixed integer linear programming. COPs are in general

More information

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs

Introduction to Mathematical Programming IE496. Final Review. Dr. Ted Ralphs Introduction to Mathematical Programming IE496 Final Review Dr. Ted Ralphs IE496 Final Review 1 Course Wrap-up: Chapter 2 In the introduction, we discussed the general framework of mathematical modeling

More information

56:272 Integer Programming & Network Flows Final Exam -- December 16, 1997

56:272 Integer Programming & Network Flows Final Exam -- December 16, 1997 56:272 Integer Programming & Network Flows Final Exam -- December 16, 1997 Answer #1 and any five of the remaining six problems! possible score 1. Multiple Choice 25 2. Traveling Salesman Problem 15 3.

More information

Solutions for Operations Research Final Exam

Solutions for Operations Research Final Exam Solutions for Operations Research Final Exam. (a) The buffer stock is B = i a i = a + a + a + a + a + a 6 + a 7 = + + + + + + =. And the transportation tableau corresponding to the transshipment problem

More information

ME 555: Distributed Optimization

ME 555: Distributed Optimization ME 555: Distributed Optimization Duke University Spring 2015 1 Administrative Course: ME 555: Distributed Optimization (Spring 2015) Instructor: Time: Location: Office hours: Website: Soomin Lee (email:

More information

Integer Programming Theory

Integer Programming Theory Integer Programming Theory Laura Galli October 24, 2016 In the following we assume all functions are linear, hence we often drop the term linear. In discrete optimization, we seek to find a solution x

More information

A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems

A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems A Feasible Region Contraction Algorithm (Frca) for Solving Linear Programming Problems E. O. Effanga Department of Mathematics/Statistics and Comp. Science University of Calabar P.M.B. 1115, Calabar, Cross

More information

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods

Unconstrained Optimization Principles of Unconstrained Optimization Search Methods 1 Nonlinear Programming Types of Nonlinear Programs (NLP) Convexity and Convex Programs NLP Solutions Unconstrained Optimization Principles of Unconstrained Optimization Search Methods Constrained Optimization

More information

A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem

A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem A Generic Separation Algorithm and Its Application to the Vehicle Routing Problem Presented by: Ted Ralphs Joint work with: Leo Kopman Les Trotter Bill Pulleyblank 1 Outline of Talk Introduction Description

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

Decomposition in Integer Linear Programming

Decomposition in Integer Linear Programming Decomposition in Integer Linear Programming T.K. Ralphs M.V. Galati December 3, 00 Abstract Both cutting plane methods and traditional decomposition methods are procedures that compute a bound on the optimal

More information

Primal Dual Schema Approach to the Labeling Problem with Applications to TSP

Primal Dual Schema Approach to the Labeling Problem with Applications to TSP 1 Primal Dual Schema Approach to the Labeling Problem with Applications to TSP Colin Brown, Simon Fraser University Instructor: Ramesh Krishnamurti The Metric Labeling Problem has many applications, especially

More information

Linear Programming Problems

Linear Programming Problems Linear Programming Problems Two common formulations of linear programming (LP) problems are: min Subject to: 1,,, 1,2,,;, max Subject to: 1,,, 1,2,,;, Linear Programming Problems The standard LP problem

More information

Introduction to Constrained Optimization

Introduction to Constrained Optimization Introduction to Constrained Optimization Duality and KKT Conditions Pratik Shah {pratik.shah [at] lnmiit.ac.in} The LNM Institute of Information Technology www.lnmiit.ac.in February 13, 2013 LNMIIT MLPR

More information

Decomposition in Integer Linear Programming

Decomposition in Integer Linear Programming Decomposition in Integer Linear Programming M.V. Galati Advanced Analytics - Operations R & D, SAS Institute, Chesterbrook, PA 9 Ted K. Ralphs Department of Industrial and System Engineering, Lehigh University,

More information