Online Learning of Assignments

Size: px
Start display at page:

Download "Online Learning of Assignments"

Transcription

1 Online Learning of Assignments Matthew Streeter Google, Inc. Pittsburgh, PA Daniel Golovin Carnegie Mellon University Pittsburgh, PA Andreas Krause California Institute of Technology Pasadena, CA Abstract Which ads should we display in sponsored search in order to maximize our revenue? How should we dynamically rank information sources to maximize the value of the ranking? These applications exhibit strong diminishing returns: Redundancy decreases the marginal utility of each ad or information source. We show that these and other problems can be formalized as repeatedly selecting an assignment of items to positions to maximize a sequence of monotone submodular functions that arrive one by one. We present an efficient algorithm for this general problem and analyze it in the no-regret model. Our algorithm possesses strong theoretical guarantees, such as a performance ratio that converges to the optimal constant of 1 1/e. We empirically evaluate our algorithm on two real-world online optimization problems on the web: ad allocation with submodular utilities, and dynamically ranking blogs to detect information cascades. 1 Introduction Consider the problem of repeatedly choosing advertisements to display in sponsored search to maximize our revenue. In this problem, there is a small set of positions on the page, and each time a query arrives we would like to assign, to each position, one out of a large number of possible ads. In this and related problems that we call online assignment learning problems, there is a set of positions, a set of items, and a sequence of rounds, and on each round we must assign an item to each position. After each round, we obtain some reward depending on the selected assignment, and we observe the value of the reward. When there is only one position, this problem becomes the well-studied multiarmed bandit problem [2]. When the positions have a linear ordering the assignment can be construed as a ranked list of elements, and the problem becomes one of selecting lists online. Online assignment learning thus models a central challenge in web search, sponsored search, news aggregators, and recommendation systems, among other applications. A common assumption made in previous work on these problems is that the quality of an assignment is the sum of a function on the (item, position pairs in the assignment. For example, online advertising models with click-through-rates [6] make an assumption of this form. More recently, there have been attempts to incorporate the value of diversity in the reward function [16]. Intuitively, even though the best K results for the query turkey might happen to be about the country, the best list of K results is likely to contain some recipes for the bird as well. This will be the case if there are diminishing returns on the number of relevant links presented to a user; for example, if it is better to present each user with at least one relevant result than to present half of the users with no relevant results and half with two relevant results. We incorporate these considerations in a flexible way by providing an algorithm that performs well whenever the reward for an assignment is a monotone submodular function of its set of (item, position pairs. Our key contributions are: (i an efficient algorithm, TABULARGREEDY, that provides a (1 1/e approximation ratio for the problem of optimizing assignments under submodular utility functions, (ii an algorithm for online learning of assignments, TGBANDIT, that has strong performance guarantees in the no-regret model, and (iii an empirical evaluation on two problems of information gathering on the web. 1

2 2 The assignment learning problem We consider problems, where we have K positions (e.g., slots for displaying ads, and need to assign to each position an item (e.g., an ad in order to maximize a utility function (e.g., the revenue from clicks on the ads. We address both the offline problem, where the utility function is specified in advance, and the online problem, where a sequence of utility functions arrives over time, and we need to repeatedly select a new assignment. The Offline Problem. In the offline problem we are given sets P 1, P 2,..., P K, where P k is the set of items that may be placed in position k. We assume without loss of generality that these sets are disjoint. 1 An assignment is a subset S V, where V = P 1 P 2 P K is the set of all items. We call an assignment feasible, if at most one item is assigned to each position (i.e., for all k, S P k 1. We use P to refer to the set of feasible assignments. Our goal is to find a feasible assignment maximizing a utility function f : 2 V R 0. As we discuss later, many important assignment problems satisfy submodularity, a natural diminishing returns property: Assigning a new item to a position k increases the utility more if few elements have been assigned yet, and less if many items have already been assigned. Formally, a utility function f is called submodular, if for all S S and s / S it holds that f(s {s} f(s f(s {s} f(s. We will also assume f is monotone (i.e., for all S S, we have f(s f(s. Our goal is thus, for a given non-negative, monotone and submodular utility function f, to find a feasible assignment S of maximum utility, S = arg max S P {f(s}. This optimization problem is NP-hard. In fact, a stronger negative result holds: Theorem 1 ([14]. For any ɛ > 0, any algorithm guaranteed to obtain a solution within a factor of (1 1/e + ɛ of max S P {f(s} requires exponentially many evaluations of f in the worst case. In light of this negative result, we can only hope to efficiently obtain a solution that achieves a fraction of (1 1/e of the optimal value. In 3.2 we develop such an algorithm. The Online Problem. The offline problem is inappropriate to model dynamic settings, where the utility function may change over time, and we need to repeatedly select new assignments, trading off exploration (experimenting with ad display to gain information about the utility function, and exploitation (displaying ads which we believe will maximize utility. More formally, we face a sequential decision problem, where, on each round (which, e.g., corresponds to a user query for a particular term, we want to select an assignment S t (ads to display. We assume that the sets P 1, P 2,..., P K are fixed in advance for all rounds. After we select the assignment we obtain reward f t (S t for some non-negative monotone submodular utility function f t. We call the setting where we do not get any information about f t beyond the reward the bandit feedback model. In contrast, in the full-information feedback model we obtain oracle access to f t (i.e., we can evaluate f t on arbitrary feasible assignments. Both models arise in real applications, as we show in 5. The goal is to maximize the total reward we obtain, namely t f t(s t. Following the multiarmed bandit literature, we evaluate our performance after T rounds by comparing our total reward against that obtained by a clairvoyant algorithm with knowledge of the sequence of functions f 1,..., f T, but with the restriction that it must select the same assignment on each round. The difference between the clairvoyant algorithm s total reward and ours is called our regret. The goal is then to develop an algorithm whose expected regret grows sublinearly in the number of rounds; such an algorithm is said to have (or be no-regret. However, since sums of submodular functions remain submodular, the clairvoyant algorithm has to solve an offline assignment problem with f(s = t f t(s. Considering Theorem 1, no polynomial-time algorithm can possibly hope to achieve a no-regret guarantee. To accommodate this fact, we discount the reward of the clairvoyant algorithm by a factor of (1 1/e: We define the (1 1/e-regret of a random sequence S 1,..., S T as ( 1 1 { T } max f t (S e S P [ T ] E f t (S t. Our goal is then to develop efficient algorithms whose (1 1/e-regret grows sublinearly in T. 1 If the same item can be placed in multiple positions, simply create multiple distinct copies of it. 2

3 Subsumed Models. Our model generalizes several common models for sponsored search ad selection, and web search results. These include models with click-through-rates, in which it is assumed that each (ad, position pair has some probability p(a, k of being clicked on, and there is some monetary reward b(a that is obtained whenever ad a is clicked on. Often, the click-throughrates are assumed to be separable, meaning p(a, k has the functional form α(a β(k for some functions α and β. See [7, 12] for more details on sponsored search ad allocation. Note that in both of these cases, the (expected reward of a set S of (ad, position pairs is (a,k S g(a, k for some nonnegative function g. It is easy to verify that such a reward function is monotone submodular. Thus, we can capture this model in our framework by setting P k = A {k}, where A is the set of ads. Another subsumed model, for web search, appears in [16]; it assumes that each user is interested in a particular set of results, and any list of results that intersects this set generates a unit of value; all other lists generate no value, and the ordering of results is irrelevant. Again, the reward function is monotone submodular. In this setting, it is desirable to display a diverse set of results in order to maximize the likelihood that at least one of them will interest the user. Our model is flexible in that we can handle position-dependent effects and diversity considerations simultaneously. For example, we can handle the case that each user u is interested in a particular set A u of ads and looks at a set I u of positions, and the reward of an assignment S is any monotoneincreasing concave function g of S (A u I u. If I u = {1, 2,..., k} and g(x = x, this models the case where the quality is the number of relevant result that appear in the first k positions. If I u equals all positions and g(x = min {x, 1} we recover the model of [16]. 3 An approximation algorithm for the offline problem 3.1 The locally greedy algorithm A simple approach to the assignment problem is the following greedy procedure: the algorithm steps through all K positions (according to some fixed, arbitrary ordering. For position k, it simply chooses the item that increases the total value as much as possible, i.e., it chooses s k = arg max s P k {f({s 1,..., s k 1 } + s}, where, for a set S and element e, we write S + e for S {e}. Perhaps surprisingly, no matter which ordering over the positions is chosen, this so-called locally greedy algorithm produces an assignment that obtains at least half the optimal value [8]. In fact, the following more general result holds. We will use this lemma in the analysis of our improved offline algorithm, which uses the locally greedy algorithm as a subroutine. Lemma 2. Suppose f : 2 V R 0 is of the form f(s = f 0 (S + K k=1 f k(s P k where f 0 : 2 V R 0 is monotone submodular, and f k : 2 P k R 0 is arbitrary for k 1. Let L be the solution returned by the locally greedy algorithm. Then f(l + f 0 (L max S P {f(s}. The proof is given in an extended version of this paper [9]. Observe that in the special case where f k 0 for all k 1, Lemma 2 says that f(l 1 2 max S P f(s. In [9] we provide a simple example showing that this 1/2 approximation ratio is tight. 3.2 An algorithm with optimal approximation ratio We now present an algorithm that achieves the optimal approximation ratio of 1 1/e, improving on the 1 2 approximation for the locally greedy algorithm. Our algorithm associates with each partition P k a color c k from a palette [C] of C colors, where we use the notation [n] = {1, 2,..., n}. For any set S V [C] and vector c = (c 1,..., c K, define sample c (S = K k=1 {x P k : (x, c k S}. Given a set S of (item, color pairs, which we may think of as labeling each item with one or more colors, sample c (S returns a set containing each item x that is labeled with whatever color c assigns to the partition that contains x. Let F (S denote the expected value of f(sample c (S when each color c k is selected uniformly at random from [C]. Our TABULARGREEDY algorithm greedily optimizes F, as shown in the following pseudocode. Observe that when C = 1, there is only one possible choice for c, and TABULARGREEDY is simply the locally greedy algorithm from 3.1. In the limit as C, TABULARGREEDY can intuitively be viewed as an algorithm for a continuous extension of the problem followed by a 3

4 Algorithm: TABULARGREEDY Input: integer C, sets P 1, P 2,..., P K, function f : 2 V R 0 (where V = K k=1 P k set G :=. for c from 1 to C do /* For each color */ for k from 1 to K do /* For each partition */ set g k,c = arg max x Pk {c} {F (G + x} /* Greedily pick g k,c */ set G := G + g k,c ; for each k [K], choose c k uniformly at random from [C]. return sample c (G, where c := (c 1,..., c K. rounding procedure, in the same spirit as Vondrák s continuous-greedy algorithm [4]. In our case, the continuous extension is to compute a probability distribution D k for each position k with support in P k (plus a special select nothing outcome, such that if we independently sample an element x k from D k, E [f({x 1,..., x K }] is maximized. It turns out that if the positions individually, greedily, and in round-robin fashion, add infinitesimal units of probability mass to their distributions so as to maximize this objective function, they achieve the same objective function value as if, rather than making decisions in a round-robin fashion, they had cooperated and added the combination of K infinitesimal probability mass units (one per position that greedily maximizes the objective function. The latter process, in turn, can be shown to be equivalent to a greedy algorithm for maximizing a (different submodular function subject to a cardinality constraint, which implies that it achieves a 1 1/e approximation ratio [15]. TABULARGREEDY represents a tradeoff between these two extremes; its performance is summarized by the following theorem. For now, we assume that the arg max in the inner loop is computed exactly. In the extended version [9], we bound the performance loss that results from approximating the arg max (e.g., by estimating F by repeated sampling. Theorem 3. Suppose f is monotone submodular. Then F (G β(k, C max S P {f(s}, where β(k, C is defined as 1 (1 1 C C ( K 2 C 1. It follows that, for any ε > 0, TABULARGREEDY achieves a (1 1/e ε approximation factor using a number of colors that is polynomial in K and 1/ε. The theorem will follow immediately from the combination of two key lemmas, which we now prove. Informally, Lemma 4 analyzes the approximation error due to the outer greedy loop of the algorithm, while Lemma 5 analyzes the approximation error due to the inner loop. Lemma 4. Let G c = {g 1,c, g 2,c,..., g K,c }, and let G c = G 1 G 2... G c 1. For each color c, choose E c R such that F (G c G c max x Rc {F (G c x} E c where R c := {R : k [K], R (P k {c} = 1} is the set of all possible choices for G c. Then where β(c = 1 ( 1 1 C C. F (G β(c max {f(s} C E c. (3.1 S P Proof (Sketch. We will refer to an element R of R c as a row, and to c as the color of the row. Let R [C] := C c=1 R c be the set of all rows. Consider the function H : 2 R [C] R 0, defined as H(R = F ( R R R. We will prove the lemma in three steps: (i H is monotone submodular, (ii TABULARGREEDY is simply the locally greedy algorithm for finding a set of C rows that maximizes H, where the c th greedy step is performed with additive error E c, and (iii TABULARGREEDY obtains the guarantee (3.1 for maximizing H, and this implies the same ratio for maximizing F. To show that H is monotone submodular, it suffices to show that F is monotone submodular. Because F (S = E c [f(sample c (S], and because a convex combination of monotone submodular functions is monotone submodular, it suffices to show that for any particular coloring c, the function f(sample c (S is monotone submodular. This follows from the definition of sample and the fact that f is monotone submodular. The second claim is true by inspection. To prove the third claim, we note that the row colors for a set of rows R can be interchanged with no effect on H(R. For problems with this special property, it is c=1 4

5 known that the locally greedy algorithm obtains an approximation ratio of β(c = 1 (1 1 C C [15]. Theorem 6 of [17] extends this result to handle additive error, and yields F (G = H({G 1, G 2,..., G C } β(c max {H(R} C E c. R R [C] : R C c=1 To complete the proof, it suffices to show that max R R[C] : R C {H(R} max S P {f(s}. This follows from the fact that for any assignment S P, we can find a set R(S of C rows such that sample c ( R R(S R = S with probability 1, and therefore H(R(S = f(s. We now bound the performance of the the inner loop of TABULARGREEDY. Lemma 5. Let f = max S P {f(s}, and let G c, G c, and R c be defined as in the statement of Lemma 4. Then, for any c [C], F (G c G c max R Rc {F (G c R} ( K 2 C 2 f. Proof (Sketch. Let N denote the number of partitions whose color (assigned by c is c. For R R c, let c (R := f(sample c (G c R f(sample c (G c, and let F c (R := F (G c R F (G c. By definition, F c (R = E c [ c (R] = P [N = 1] E c [ c (R N = 1] + P [N 2] E c [ c (R N 2], where we have used the fact that c (R = 0 when N = 0. The idea of the proof is that the first of these terms dominates as C, and that E c [ c (R N = 1] can be optimized exactly simply by optimizing each element of P k {c} independently. Specifically, it can be seen that E c [ c (R N = 1] = K k=1 f k(r (P k {c} for suitable f k. Additionally, f 0 (R = P [N 2] E c [ c (R N 2] is a monotone submodular function of a set of (item, color pairs, for the same reasons F is. Applying Lemma 2 with these {f k : k 0} yields F c (G c + P [N 2] E c [ c (G c N 2] max R Rc {F c (R}. To complete the proof, it suffices to show P [N 2] ( K 2 C 2 and E c [ c (G c N 2] f. The first inequality holds because, if we let M be the number of pairs of partitions that are both assigned color c, we have P [N 2] = P [M 1] E [M] = ( K 2 C 2. The second inequality follows from the fact that for any c we have c (G c f(sample c (G c G c f. 4 An algorithm for online learning of assignments We now transform the offline algorithm of 3.2 into an online algorithm. The high-level idea behind this transformation is to replace each greedy decision made by the offline algorithm with a no-regret online algorithm. A similar approach was used in [16] and [18] to obtain an online algorithm for different (simpler online problems. Algorithm: TGBANDIT (described in the full-information feedback model Input: integer C, sets P 1, P 2,..., P K for each k [K] and c [C], let E k,c be a no-regret algorithm with action set P k {c}. for t from 1 to T do for each k [K] and c [C], let gk,c t P k {c} be the action selected by E k,c for each k [K], choose c k ({ uniformly at random from } [C]. Define c = (c 1,..., c K. select the set G t = sample c gk,c t : k [K], c [C] observe f t, and let F t (S := f t (sample c (S for each k [K], c [C] do } { } define G t k,c {g k t,c : k [K], c < c gk t,c : k < k for each x P k {c}, feed back F t (G t k,c + x to E k,c as the reward for choosing x The following theorem summarizes the performance of TGBANDIT. Theorem 6. Let r k,c be the regret of E k,c, and let β(k, C = 1 ( 1 1 C ( C K 2 C 1. Then [ T ] { T } [ K ] C E f t (G t β(k, C max f t (S E r k,c. S P k=1 c=1 5

6 Observe that Theorem 6 is similar to Theorem 3, with the addition of the E [r k,c ] terms. The idea of the proof is to view TGBANDIT as a version of TABULARGREEDY that, instead of greedily selecting single (element,color pairs g k,c P k {c}, greedily selects (element vector, color pairs g k,c P T k {c} (here, P T k is the T th power of the set P k. We allow for the case that the greedy decision is made imperfectly, with additive error r k,c ; this is the source of the extra terms. Once this correspondence is established, the theorem follows along the lines of Theorem 3. For a proof, see the extended version [9]. Corollary 7. If TGBANDIT is run with randomized weighted majority [5] as the subroutine, then [ T ] { T } ( K E f t (G t β(k, C max f t (S O C T log Pk. S P where β(k, C = 1 ( 1 1 C C ( K 2 C 1. Optimizing for C in Corollary 7 { yields (1 1 e -regret Θ(K 3/2 T 1/4 OPT ignoring logarithmic T } factors, where OPT := max S P f t(s is the value of the static optimum. Dealing with bandit feedback. TGBANDIT can be modified to work in the bandit feedback model. The idea behind this modification is that on each round we explore with some small probability, in such a way that on each round we obtain an unbiased estimate of the desired feedback values F t (G t k,c + x for each k [K], c [C], and x P k. This technique can be used ( to achieve a bound similar to the one stated in Corollary 7, but with an additive regret term of O (T V CK 2 3 (log V 1 3. Stronger notions of regret. By substituting in different algorithms for the subroutines E k,c, we can obtain additional guarantees. For example, Blum and Mansour [3] consider online problems in which we are given time-selection functions I 1, I 2,..., I M. Each time-selection function I : [T ] [0, 1] associates a weight with each round, and defines a corresponding weighted notion of regret in the natural way. Blum and Mansour s algorithm guarantees low weighted regret with respect to all M time selection functions simultaneously. This can be used to obtain low regret with respect to different (possibly overlapping windows of time simultaneously, or to obtain low regret with respect to subsets of rounds that have particular features. By using their algorithm as a subroutine within TGBANDIT, we get similar guarantees, both in the full information and bandit feedback models. 5 Evaluation We evaluate TGBANDIT experimentally on two applications: Learning to rank blogs that are effective in detecting cascades of information, and allocating advertisements to maximize revenue. 5.1 Online learning of diverse blog rankings We consider the problem of ranking a set of blogs and news sources on the web. Our approach is based on the following idea: A blogger writes a posting, and, after some time, other postings link to it, forming cascades of information propagating through the network of blogs. More formally, an information cascade is a directed acyclic graph of vertices (each vertex corresponds to a posting at some blog, where edges are annotated by the time difference between the postings. Based on this notion of an information cascade, we would like to select blogs that detect big cascades (containing many nodes as early as possible (i.e., we want to learn about an important event before most other readers. In [13] it is shown how one can formalize this notion of utility using a monotone submodular function that measures the informativeness of a subset of blogs. Optimizing the submodular function yields a small set of blogs that covers most cascades. This utility function prefers diverse sets of blogs, minimizing the overlap of the detected cascades, and therefore minimizing redundancy. The work by [13] leaves two major shortcomings: Firstly, they select a set of blogs rather than a ranking, which is of practical importance for the presentation on a web service. Secondly, they do not address the problem of sequential prediction, where the set of blogs must be updated dynamically over time. In this paper, we address these shortcomings. 6 k=1

7 Performance 8 x 104 Average Maximum Number of colors (a Blogs: Offline results Avg. normalized performance colors 1 color Number of rounds (days (b Blogs: Online results Average Payoff colors 2 colors 1 color Number of rounds (c Ad display: Online results Figure 1: (a,b Results for discounted blog ranking (γ = 0.8, in offline (a and online (b setting. (c Performance of TGBANDIT with C = 1, 2, and 4 colors for the sponsored search ad selection problem (each round is a query. Note that C = 1 corresponds to the online algorithm of [16, 18]. Results on offline blog ranking. In order to model the blog ranking problem, we adopt the assumption that different users have different attention spans: Each user will only consider blogs appearing in a particular subset of positions. In our experiments, we assume that the probability that a user is willing to look at position k is proportional to γ k, for some discount factor 0 < γ < 1. More formally, let g be the monotone submodular function measuring the informativeness of any set of blogs, defined as in [13]. Let P k = B {k}, where B is the set of blogs. Given an assignment S P, let S [k] = S {P 1 P 2... P k } be the assignment of blogs to positions 1 through k. We define the discounted value of the assignment S as f(s = K ( k=1 γk g(s [k] g(s [k 1]. It can be seen that f : 2 V R 0 is monotone submodular. For our experiments, we use the data set of [13], consisting of 45,192 blogs, 16,551 cascades, and 2 million postings collected during 12 months of We use the population affected objective of [13], and use a discount factor of γ = 0.8. Based on this data, we run our TABULARGREEDY algorithm with varying numbers of colors C on the blog data set. Fig. 1(a presents the results of this experiment. For each value of C, we generate 200 rankings, and report both the average performance and the maximum performance over the 200 trials. Increasing C leads to an improved performance over the locally greedy algorithm (C = 1. Results on online learning of blog rankings. We now consider the online problem where on each round t we want to output a ranking S t. After we select the ranking, a new set of cascades occurs, modeled using a separate submodular function f t, and we obtain a reward of f t (S t. In our experiments, we choose one assignment per day, and define f t as the utility associated with the cascades occurring on that day. Note that f t allows us to evaluate the performance of any possible ranking S t, hence we can apply TGBANDIT in the full-information feedback model. We compare the performance of our online algorithm using C = 1 and C = 4. Fig. 1(b presents the average cumulative reward gained over time by both algorithms. We normalize the average reward by the utility achieved by the TABULARGREEDY algorithm (with C = 1 applied to the entire data set. Fig. 1(b shows that the performance of both algorithms rapidly (within the first 47 rounds converges to the performance of the offline algorithm. The TGBANDIT algorithm with C = 4 levels out at an approximately 4% higher reward than the algorithm with C = Online ad display We evaluate TGBANDIT for the sponsored search ad selection problem in a simple Markovian model incorporating the value of diverse results and complex position-dependence among clicks. In this model, each user u is defined by two sets of probabilities: p click (a for each ad a A, and p abandon (k for each position k [K]. When presented an assignment of ads {a 1, a 2,..., a K }, where a k occupies position k, the user scans the positions in increasing order. For each position k, the user clicks on a k with probability p click (a k, leaving the results page forever. Otherwise, with probability (1 p click (a k p abandon (k, the user loses interest and abandons the results without clicking on anything. Finally, with probability (1 p click (a k (1 p abandon (k, the user proceeds to look at position k + 1. The reward function f t is the number of clicks, which is either zero or one. We only receive information about f t (S t (i.e., bandit feedback. 7

8 In our evaluation, there are 5 positions, 20 available ads, and two (equally frequent types of users: type 1 users interested in all positions (p abandon 0, and type 2 users that quickly lose interest (p abandon 0.5. There are also two types of ads, half of type 1 and half of type 2, and users are probabilistically more interested in ads of their own type than those of the opposite type. Specifically, for both types of users we set p click (a = 0.5 if a has the same type as the user, and p click (a = 0.2 otherwise. In Fig. 1(c we compare the performance of TGBANDIT with C = 4 to the online algorithm of [16, 18], based on the average of 100 experiments. The latter algorithm is equivalent to running TGBANDIT with C = 1. They perform similarly in the first 10 4 rounds; thereafter the former algorithm dominates. It can be shown that with several different types of users with distinct p click ( functions the offline problem of finding an assignment within 1 1 e + ε of optimal is NP-hard. This is in contrast to the case in which p click and p abandon are the same for all users; in this case the offline problem simply requires finding an optimal policy for a Markov decision process, which can be done efficiently using well-known algorithms. A slightly different Markov model of user behavior which is efficiently solvable was considered in [1]. In that model, p click and p abandon are the same for all users, and p abandon is a function of the ad in the slot currently being scanned rather than its index. 6 Related Work For a general introduction to the literature on submodular function maximization, see [19]. For applications of submodularity to machine learning and AI see [11]. Our offline problem is known as maximizing a monotone submodular function subject to a (simple partition matroid constraint in the operations research and theoretical computer science communities. The study of this problem culminated in the elegant (1 1/e approximation algorithm of Vondrák [20] and a matching unconditional lower bound of Mirrokni et al. [14]. Vondrák s algorithm, called the continuous-greedy algorithm, has also been extended to handle arbitrary matroid constraints [4]. The continuous-greedy algorithm, however, cannot be applied to our problem directly, because it requires the ability to sample f( on infeasible sets S / P. In our context, this means it must have the ability to ask (for example what the revenue will be if ads a 1 and a 2 are placed in position #1 simultaneously. We do not know how to answer such questions in a way that leads to meaningful performance guarantees. In the online setting, the most closely related work is that of Streeter and Golovin [18]. Like us, they consider sequences of monotone submodular reward functions that arrive online, and develop an online algorithm that uses multi-armed bandit algorithms as subroutines. The key difference from our work is that, as in [16], they are concerned with selecting a set of K items rather than the more general problem of selecting an assignment of items to positions addressed in this paper. Kakade et al. [10] considered the general problem of using α-approximation algorithms to construct no α-regret online algorithms, and essentially proved it could be done for the class of linear optimization problems in which the cost function has the form c(s, w for a solution S and weight vector w, and c(s, w is linear in w. However, their result is orthogonal to ours, because our objective function is submodular and not linear 2. 7 Conclusions In this paper, we showed that important problems, such as ad display in sponsored search and computing diverse rankings of information sources on the web, require optimizing assignments under submodular utility functions. We developed an efficient algorithm, TABULARGREEDY, which obtains the optimal approximation ratio of (1 1/e for this NP-hard optimization problem. We also developed an online algorithm, TGBANDIT, that asymptotically achieves no (1 1/e-regret for the problem of repeatedly selecting informative assignments, under the full-information and bandit-feedback settings. Finally, we demonstrated that our algorithm outperforms previous work on two real world problems, namely online ranking of informative blogs and ad allocation. Acknowledgments. This work was supported in part by Microsoft Corporation through a gift as well as through the Center for Computational Thinking at Carnegie Mellon, by NSF ITR grant CCR (The Aladdin Center, and by ONR grant N One may linearize a submodular function by using a separate dimension for every possible function argument, but this leads to exponentially worse convergence time and regret bounds for the algorithms in [10] relative to TGBANDIT. 8

9 References [1] Gagan Aggarwal, Jon Feldman, S. Muthukrishnan, and Martin Pál. Sponsored search auctions with markovian users. In WINE, pages , [2] Peter Auer, Nicolò Cesa-Bianchi, Yoav Freund, and Robert E. Schapire. The nonstochastic multiarmed bandit problem. SIAM Journal on Computing, 32(1:48 77, [3] Avrim Blum and Yishay Mansour. From external to internal regret. Journal of Machine Learning Research, 8: , [4] Gruia Calinescu, Chandra Chekuri, Martin Pál, and Jan Vondrák. Maximizing a submodular set function subject to a matroid constraint. SIAM Journal on Computing. To appear. [5] Nicolò Cesa-Bianchi, Yoav Freund, David Haussler, David P. Helmbold, Robert E. Schapire, and Manfred K. Warmuth. How to use expert advice. J. ACM, 44(3: , [6] Benjamin Edelman, Michael Ostrovsky, and Michael Schwarz. Internet advertising and the generalized second price auction: Selling billions of dollars worth of keywords. American Economic Review, 97(1: , [7] Jon Feldman and S. Muthukrishnan. Algorithmic methods for sponsored search advertising. In Zhen Liu and Cathy H. Xia, editors, Performance Modeling and Engineering [8] Marshall L. Fisher, George L. Nemhauser, and Laurence A. Wolsey. An analysis of approximations for maximizing submodular set functions - II. Mathematical Programming Study, (8:73 87, [9] Daniel Golovin, Andreas Krause, and Matthew Streeter. Online learning of assignments that maximize submodular functions. CoRR, abs/ , [10] Sham M. Kakade, Adam Tauman Kalai, and Katrina Ligett. Playing games with approximation algorithms. In STOC, pages , [11] Andreas Krause and Carlos Guestrin. Beyond convexity: Submodularity in machine learning. Tutorial at ICML [12] Sébastien Lahaie, David M. Pennock, Amin Saberi, and Rakesh V. Vohra. Sponsored search auctions. In Noam Nisan, Tim Roughgarden, Eva Tardos, and Vijay V. Vazirani, editors, Algorithmic Game Theory. Cambridge University Press, New York, NY, USA, [13] Jure Leskovec, Andreas Krause, Carlos Guestrin, Christos Faloutsos, Jeanne VanBriesen, and Natalie Glance. Cost-effective outbreak detection in networks. In KDD, pages , [14] Vahab Mirrokni, Michael Schapira, and Jan Vondrák. Tight information-theoretic lower bounds for welfare maximization in combinatorial auctions. In EC, pages 70 77, [15] George L. Nemhauser, Laurence A. Wolsey, and Marshall L. Fisher. An analysis of approximations for maximizing submodular set functions - I. Mathematical Programming, 14(1: , [16] Filip Radlinski, Robert Kleinberg, and Thorsten Joachims. Learning diverse rankings with multi-armed bandits. In ICML, pages , [17] Matthew Streeter and Daniel Golovin. An online algorithm for maximizing submodular functions. Technical Report CMU-CS , Carnegie Mellon University, [18] Matthew Streeter and Daniel Golovin. An online algorithm for maximizing submodular functions. In NIPS, pages , [19] Jan Vondrák. Submodularity in Combinatorial Optimization. PhD thesis, Charles University, Prague, Czech Republic, [20] Jan Vondrák. Optimal approximation for the submodular welfare problem in the value oracle model. In STOC, pages 67 74,

CS 598CSC: Approximation Algorithms Lecture date: March 2, 2011 Instructor: Chandra Chekuri

CS 598CSC: Approximation Algorithms Lecture date: March 2, 2011 Instructor: Chandra Chekuri CS 598CSC: Approximation Algorithms Lecture date: March, 011 Instructor: Chandra Chekuri Scribe: CC Local search is a powerful and widely used heuristic method (with various extensions). In this lecture

More information

Viral Marketing and Outbreak Detection. Fang Jin Yao Zhang

Viral Marketing and Outbreak Detection. Fang Jin Yao Zhang Viral Marketing and Outbreak Detection Fang Jin Yao Zhang Paper 1: Maximizing the Spread of Influence through a Social Network Authors: David Kempe, Jon Kleinberg, Éva Tardos KDD 2003 Outline Problem Statement

More information

Communication Complexity of Combinatorial Auctions with Submodular Valuations

Communication Complexity of Combinatorial Auctions with Submodular Valuations Communication Complexity of Combinatorial Auctions with Submodular Valuations Shahar Dobzinski Weizmann Institute of Science Jan Vondrák IBM Almaden Research ACM-SIAM SODA 2013, New Orleans Dobzinski-Vondrák

More information

Online Stochastic Matching CMSC 858F: Algorithmic Game Theory Fall 2010

Online Stochastic Matching CMSC 858F: Algorithmic Game Theory Fall 2010 Online Stochastic Matching CMSC 858F: Algorithmic Game Theory Fall 2010 Barna Saha, Vahid Liaghat Abstract This summary is mostly based on the work of Saberi et al. [1] on online stochastic matching problem

More information

Approximation Algorithms for Item Pricing

Approximation Algorithms for Item Pricing Approximation Algorithms for Item Pricing Maria-Florina Balcan July 2005 CMU-CS-05-176 Avrim Blum School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 School of Computer Science,

More information

Jure Leskovec Machine Learning Department Carnegie Mellon University

Jure Leskovec Machine Learning Department Carnegie Mellon University Jure Leskovec Machine Learning Department Carnegie Mellon University Currently: Soon: Today: Large on line systems have detailed records of human activity On line communities: Facebook (64 million users,

More information

Online Bipartite Matching: A Survey and A New Problem

Online Bipartite Matching: A Survey and A New Problem Online Bipartite Matching: A Survey and A New Problem Xixuan Feng xfeng@cs.wisc.edu Abstract We study the problem of online bipartite matching, where algorithms have to draw irrevocable matchings based

More information

Improved Approximation Algorithms for Budgeted Allocations

Improved Approximation Algorithms for Budgeted Allocations Improved Approximation Algorithms for Budgeted Allocations Yossi Azar 1,BenjaminBirnbaum 2, Anna R. Karlin 2, Claire Mathieu 3, and C. Thach Nguyen 2 1 Microsoft Research and Tel-Aviv University azar@tau.ac.il

More information

The Encoding Complexity of Network Coding

The Encoding Complexity of Network Coding The Encoding Complexity of Network Coding Michael Langberg Alexander Sprintson Jehoshua Bruck California Institute of Technology Email: mikel,spalex,bruck @caltech.edu Abstract In the multicast network

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A 4 credit unit course Part of Theoretical Computer Science courses at the Laboratory of Mathematics There will be 4 hours

More information

Lecture 9: (Semi-)bandits and experts with linear costs (part I)

Lecture 9: (Semi-)bandits and experts with linear costs (part I) CMSC 858G: Bandits, Experts and Games 11/03/16 Lecture 9: (Semi-)bandits and experts with linear costs (part I) Instructor: Alex Slivkins Scribed by: Amr Sharaf In this lecture, we will study bandit problems

More information

CS224W: Analysis of Networks Jure Leskovec, Stanford University

CS224W: Analysis of Networks Jure Leskovec, Stanford University HW2 Q1.1 parts (b) and (c) cancelled. HW3 released. It is long. Start early. CS224W: Analysis of Networks Jure Leskovec, Stanford University http://cs224w.stanford.edu 10/26/17 Jure Leskovec, Stanford

More information

6. Lecture notes on matroid intersection

6. Lecture notes on matroid intersection Massachusetts Institute of Technology 18.453: Combinatorial Optimization Michel X. Goemans May 2, 2017 6. Lecture notes on matroid intersection One nice feature about matroids is that a simple greedy algorithm

More information

Parameterized graph separation problems

Parameterized graph separation problems Parameterized graph separation problems Dániel Marx Department of Computer Science and Information Theory, Budapest University of Technology and Economics Budapest, H-1521, Hungary, dmarx@cs.bme.hu Abstract.

More information

1 The Traveling Salesperson Problem (TSP)

1 The Traveling Salesperson Problem (TSP) CS 598CSC: Approximation Algorithms Lecture date: January 23, 2009 Instructor: Chandra Chekuri Scribe: Sungjin Im In the previous lecture, we had a quick overview of several basic aspects of approximation

More information

A survey of submodular functions maximization. Yao Zhang 03/19/2015

A survey of submodular functions maximization. Yao Zhang 03/19/2015 A survey of submodular functions maximization Yao Zhang 03/19/2015 Example Deploy sensors in the water distribution network to detect contamination F(S): the performance of the detection when a set S of

More information

5. Lecture notes on matroid intersection

5. Lecture notes on matroid intersection Massachusetts Institute of Technology Handout 14 18.433: Combinatorial Optimization April 1st, 2009 Michel X. Goemans 5. Lecture notes on matroid intersection One nice feature about matroids is that a

More information

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 The Encoding Complexity of Network Coding Michael Langberg, Member, IEEE, Alexander Sprintson, Member, IEEE, and Jehoshua Bruck,

More information

Theorem 2.9: nearest addition algorithm

Theorem 2.9: nearest addition algorithm There are severe limits on our ability to compute near-optimal tours It is NP-complete to decide whether a given undirected =(,)has a Hamiltonian cycle An approximation algorithm for the TSP can be used

More information

Lecture 5: Duality Theory

Lecture 5: Duality Theory Lecture 5: Duality Theory Rajat Mittal IIT Kanpur The objective of this lecture note will be to learn duality theory of linear programming. We are planning to answer following questions. What are hyperplane

More information

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs Graphs and Network Flows IE411 Lecture 21 Dr. Ted Ralphs IE411 Lecture 21 1 Combinatorial Optimization and Network Flows In general, most combinatorial optimization and integer programming problems are

More information

CS261: A Second Course in Algorithms Lecture #16: The Traveling Salesman Problem

CS261: A Second Course in Algorithms Lecture #16: The Traveling Salesman Problem CS61: A Second Course in Algorithms Lecture #16: The Traveling Salesman Problem Tim Roughgarden February 5, 016 1 The Traveling Salesman Problem (TSP) In this lecture we study a famous computational problem,

More information

Polynomial-Time Approximation Algorithms

Polynomial-Time Approximation Algorithms 6.854 Advanced Algorithms Lecture 20: 10/27/2006 Lecturer: David Karger Scribes: Matt Doherty, John Nham, Sergiy Sidenko, David Schultz Polynomial-Time Approximation Algorithms NP-hard problems are a vast

More information

On the Computational Complexity of Nash Equilibria for (0, 1) Bimatrix Games

On the Computational Complexity of Nash Equilibria for (0, 1) Bimatrix Games On the Computational Complexity of Nash Equilibria for (0, 1) Bimatrix Games Bruno Codenotti Daniel Štefankovič Abstract The computational complexity of finding a Nash equilibrium in a nonzero sum bimatrix

More information

On Distributed Submodular Maximization with Limited Information

On Distributed Submodular Maximization with Limited Information On Distributed Submodular Maximization with Limited Information Bahman Gharesifard Stephen L. Smith Abstract This paper considers a class of distributed submodular maximization problems in which each agent

More information

Interleaving Schemes on Circulant Graphs with Two Offsets

Interleaving Schemes on Circulant Graphs with Two Offsets Interleaving Schemes on Circulant raphs with Two Offsets Aleksandrs Slivkins Department of Computer Science Cornell University Ithaca, NY 14853 slivkins@cs.cornell.edu Jehoshua Bruck Department of Electrical

More information

Topic: Local Search: Max-Cut, Facility Location Date: 2/13/2007

Topic: Local Search: Max-Cut, Facility Location Date: 2/13/2007 CS880: Approximations Algorithms Scribe: Chi Man Liu Lecturer: Shuchi Chawla Topic: Local Search: Max-Cut, Facility Location Date: 2/3/2007 In previous lectures we saw how dynamic programming could be

More information

Math 5593 Linear Programming Lecture Notes

Math 5593 Linear Programming Lecture Notes Math 5593 Linear Programming Lecture Notes Unit II: Theory & Foundations (Convex Analysis) University of Colorado Denver, Fall 2013 Topics 1 Convex Sets 1 1.1 Basic Properties (Luenberger-Ye Appendix B.1).........................

More information

Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems

Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems Dual-Based Approximation Algorithms for Cut-Based Network Connectivity Problems Benjamin Grimmer bdg79@cornell.edu arxiv:1508.05567v2 [cs.ds] 20 Jul 2017 Abstract We consider a variety of NP-Complete network

More information

5 MST and Greedy Algorithms

5 MST and Greedy Algorithms 5 MST and Greedy Algorithms One of the traditional and practically motivated problems of discrete optimization asks for a minimal interconnection of a given set of terminals (meaning that every pair will

More information

Copyright 2000, Kevin Wayne 1

Copyright 2000, Kevin Wayne 1 Guessing Game: NP-Complete? 1. LONGEST-PATH: Given a graph G = (V, E), does there exists a simple path of length at least k edges? YES. SHORTEST-PATH: Given a graph G = (V, E), does there exists a simple

More information

Online Facility Location

Online Facility Location Online Facility Location Adam Meyerson Abstract We consider the online variant of facility location, in which demand points arrive one at a time and we must maintain a set of facilities to service these

More information

Stanford University CS261: Optimization Handout 1 Luca Trevisan January 4, 2011

Stanford University CS261: Optimization Handout 1 Luca Trevisan January 4, 2011 Stanford University CS261: Optimization Handout 1 Luca Trevisan January 4, 2011 Lecture 1 In which we describe what this course is about and give two simple examples of approximation algorithms 1 Overview

More information

On Covering a Graph Optimally with Induced Subgraphs

On Covering a Graph Optimally with Induced Subgraphs On Covering a Graph Optimally with Induced Subgraphs Shripad Thite April 1, 006 Abstract We consider the problem of covering a graph with a given number of induced subgraphs so that the maximum number

More information

Approximation Algorithms

Approximation Algorithms Approximation Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

PTAS for Matroid Matching

PTAS for Matroid Matching PTAS for Matroid Matching Jon Lee 1 Maxim Sviridenko 1 Jan Vondrák 2 1 IBM Watson Research Center Yorktown Heights, NY 2 IBM Almaden Research Center San Jose, CA May 6, 2010 Jan Vondrák (IBM Almaden) PTAS

More information

Monotone Paths in Geometric Triangulations

Monotone Paths in Geometric Triangulations Monotone Paths in Geometric Triangulations Adrian Dumitrescu Ritankar Mandal Csaba D. Tóth November 19, 2017 Abstract (I) We prove that the (maximum) number of monotone paths in a geometric triangulation

More information

Lecture 2. 1 Introduction. 2 The Set Cover Problem. COMPSCI 632: Approximation Algorithms August 30, 2017

Lecture 2. 1 Introduction. 2 The Set Cover Problem. COMPSCI 632: Approximation Algorithms August 30, 2017 COMPSCI 632: Approximation Algorithms August 30, 2017 Lecturer: Debmalya Panigrahi Lecture 2 Scribe: Nat Kell 1 Introduction In this lecture, we examine a variety of problems for which we give greedy approximation

More information

5 MST and Greedy Algorithms

5 MST and Greedy Algorithms 5 MST and Greedy Algorithms One of the traditional and practically motivated problems of discrete optimization asks for a minimal interconnection of a given set of terminals (meaning that every pair will

More information

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T.

Formal Model. Figure 1: The target concept T is a subset of the concept S = [0, 1]. The search agent needs to search S for a point in T. Although this paper analyzes shaping with respect to its benefits on search problems, the reader should recognize that shaping is often intimately related to reinforcement learning. The objective in reinforcement

More information

Network Improvement for Equilibrium Routing

Network Improvement for Equilibrium Routing Network Improvement for Equilibrium Routing UMANG BHASKAR University of Waterloo and KATRINA LIGETT California Institute of Technology Routing games are frequently used to model the behavior of traffic

More information

Impersonation-Based Mechanisms

Impersonation-Based Mechanisms Impersonation-Based Mechanisms Moshe Babaioff, Ron Lavi, and Elan Pavlov Abstract In this paper we present a general scheme to create mechanisms that approximate the social welfare in the presence of selfish

More information

Secretary Problems and Incentives via Linear Programming

Secretary Problems and Incentives via Linear Programming Secretary Problems and Incentives via Linear Programming Niv Buchbinder Microsoft Research, New England and Kamal Jain Microsoft Research, Redmond and Mohit Singh Microsoft Research, New England In the

More information

On the Max Coloring Problem

On the Max Coloring Problem On the Max Coloring Problem Leah Epstein Asaf Levin May 22, 2010 Abstract We consider max coloring on hereditary graph classes. The problem is defined as follows. Given a graph G = (V, E) and positive

More information

8 Matroid Intersection

8 Matroid Intersection 8 Matroid Intersection 8.1 Definition and examples 8.2 Matroid Intersection Algorithm 8.1 Definitions Given two matroids M 1 = (X, I 1 ) and M 2 = (X, I 2 ) on the same set X, their intersection is M 1

More information

The Index Coding Problem: A Game-Theoretical Perspective

The Index Coding Problem: A Game-Theoretical Perspective The Index Coding Problem: A Game-Theoretical Perspective Yu-Pin Hsu, I-Hong Hou, and Alex Sprintson Department of Electrical and Computer Engineering Texas A&M University {yupinhsu, ihou, spalex}@tamu.edu

More information

Discrete Optimization. Lecture Notes 2

Discrete Optimization. Lecture Notes 2 Discrete Optimization. Lecture Notes 2 Disjunctive Constraints Defining variables and formulating linear constraints can be straightforward or more sophisticated, depending on the problem structure. The

More information

Crossing Families. Abstract

Crossing Families. Abstract Crossing Families Boris Aronov 1, Paul Erdős 2, Wayne Goddard 3, Daniel J. Kleitman 3, Michael Klugerman 3, János Pach 2,4, Leonard J. Schulman 3 Abstract Given a set of points in the plane, a crossing

More information

A Note on Scheduling Parallel Unit Jobs on Hypercubes

A Note on Scheduling Parallel Unit Jobs on Hypercubes A Note on Scheduling Parallel Unit Jobs on Hypercubes Ondřej Zajíček Abstract We study the problem of scheduling independent unit-time parallel jobs on hypercubes. A parallel job has to be scheduled between

More information

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/3/15

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/3/15 600.363 Introduction to Algorithms / 600.463 Algorithms I Lecturer: Michael Dinitz Topic: Algorithms and Game Theory Date: 12/3/15 25.1 Introduction Today we re going to spend some time discussing game

More information

Lecture 2 - Introduction to Polytopes

Lecture 2 - Introduction to Polytopes Lecture 2 - Introduction to Polytopes Optimization and Approximation - ENS M1 Nicolas Bousquet 1 Reminder of Linear Algebra definitions Let x 1,..., x m be points in R n and λ 1,..., λ m be real numbers.

More information

Learning Socially Optimal Information Systems from Egoistic Users

Learning Socially Optimal Information Systems from Egoistic Users Learning Socially Optimal Information Systems from Egoistic Users Karthik Raman Thorsten Joachims Department of Computer Science Cornell University, Ithaca NY karthik@cs.cornell.edu www.cs.cornell.edu/

More information

1 Overview. 2 Applications of submodular maximization. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 Applications of submodular maximization. AM 221: Advanced Optimization Spring 2016 AM : Advanced Optimization Spring 06 Prof. Yaron Singer Lecture 0 April th Overview Last time we saw the problem of Combinatorial Auctions and framed it as a submodular maximization problem under a partition

More information

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Algorithms For Inference Fall 2014 Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 Recitation-6: Hardness of Inference Contents 1 NP-Hardness Part-II

More information

Online Ad Allocation: Theory and Practice

Online Ad Allocation: Theory and Practice Online Ad Allocation: Theory and Practice Vahab Mirrokni December 19, 2014 Based on recent papers in collaboration with my colleagues at Google, Columbia Univ., Cornell, MIT, and Stanford. [S. Balseiro,

More information

Exact Algorithms Lecture 7: FPT Hardness and the ETH

Exact Algorithms Lecture 7: FPT Hardness and the ETH Exact Algorithms Lecture 7: FPT Hardness and the ETH February 12, 2016 Lecturer: Michael Lampis 1 Reminder: FPT algorithms Definition 1. A parameterized problem is a function from (χ, k) {0, 1} N to {0,

More information

Constant Price of Anarchy in Network Creation Games via Public Service Advertising

Constant Price of Anarchy in Network Creation Games via Public Service Advertising Constant Price of Anarchy in Network Creation Games via Public Service Advertising Erik D. Demaine and Morteza Zadimoghaddam MIT Computer Science and Artificial Intelligence Laboratory, Cambridge, MA 02139,

More information

Fast Convergence of Regularized Learning in Games

Fast Convergence of Regularized Learning in Games Fast Convergence of Regularized Learning in Games Vasilis Syrgkanis Alekh Agarwal Haipeng Luo Robert Schapire Microsoft Research NYC Microsoft Research NYC Princeton University Microsoft Research NYC Strategic

More information

Greedy Algorithms 1 {K(S) K(S) C} For large values of d, brute force search is not feasible because there are 2 d {1,..., d}.

Greedy Algorithms 1 {K(S) K(S) C} For large values of d, brute force search is not feasible because there are 2 d {1,..., d}. Greedy Algorithms 1 Simple Knapsack Problem Greedy Algorithms form an important class of algorithmic techniques. We illustrate the idea by applying it to a simplified version of the Knapsack Problem. Informally,

More information

Algorithms for Euclidean TSP

Algorithms for Euclidean TSP This week, paper [2] by Arora. See the slides for figures. See also http://www.cs.princeton.edu/~arora/pubs/arorageo.ps Algorithms for Introduction This lecture is about the polynomial time approximation

More information

Fall CS598CC: Approximation Algorithms. Chandra Chekuri

Fall CS598CC: Approximation Algorithms. Chandra Chekuri Fall 2006 CS598CC: Approximation Algorithms Chandra Chekuri Administrivia http://www.cs.uiuc.edu/homes/chekuri/teaching/fall2006/approx.htm Grading: 4 home works (60-70%), 1 take home final (30-40%) Mailing

More information

Core Membership Computation for Succinct Representations of Coalitional Games

Core Membership Computation for Succinct Representations of Coalitional Games Core Membership Computation for Succinct Representations of Coalitional Games Xi Alice Gao May 11, 2009 Abstract In this paper, I compare and contrast two formal results on the computational complexity

More information

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg and Eva Tardos

Maximizing the Spread of Influence through a Social Network. David Kempe, Jon Kleinberg and Eva Tardos Maximizing the Spread of Influence through a Social Network David Kempe, Jon Kleinberg and Eva Tardos Group 9 Lauren Thomas, Ryan Lieblein, Joshua Hammock and Mary Hanvey Introduction In a social network,

More information

A General Class of Heuristics for Minimum Weight Perfect Matching and Fast Special Cases with Doubly and Triply Logarithmic Errors 1

A General Class of Heuristics for Minimum Weight Perfect Matching and Fast Special Cases with Doubly and Triply Logarithmic Errors 1 Algorithmica (1997) 18: 544 559 Algorithmica 1997 Springer-Verlag New York Inc. A General Class of Heuristics for Minimum Weight Perfect Matching and Fast Special Cases with Doubly and Triply Logarithmic

More information

Submodular Optimization in Computational Sustainability. Andreas Krause

Submodular Optimization in Computational Sustainability. Andreas Krause Submodular Optimization in Computational Sustainability Andreas Krause Master Class at CompSust 2012 Combinatorial optimization in computational sustainability Many applications in computational sustainability

More information

A Network Coloring Game

A Network Coloring Game A Network Coloring Game Kamalika Chaudhuri, Fan Chung 2, and Mohammad Shoaib Jamall 2 Information Theory and Applications Center, UC San Diego kamalika@soe.ucsd.edu 2 Department of Mathematics, UC San

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Approximation algorithms Date: 11/27/18 22.1 Introduction We spent the last two lectures proving that for certain problems, we can

More information

Information Filtering for arxiv.org

Information Filtering for arxiv.org Information Filtering for arxiv.org Bandits, Exploration vs. Exploitation, & the Cold Start Problem Peter Frazier School of Operations Research & Information Engineering Cornell University with Xiaoting

More information

Notes for Lecture 24

Notes for Lecture 24 U.C. Berkeley CS170: Intro to CS Theory Handout N24 Professor Luca Trevisan December 4, 2001 Notes for Lecture 24 1 Some NP-complete Numerical Problems 1.1 Subset Sum The Subset Sum problem is defined

More information

Online Distributed Sensor Selection

Online Distributed Sensor Selection Online Distributed Sensor Selection Daniel Golovin Caltech Matthew Faulkner Caltech Andreas Krause Caltech ABSTRACT A key problem in sensor networks is to decide which sensors to query when, in order to

More information

Coverage Approximation Algorithms

Coverage Approximation Algorithms DATA MINING LECTURE 12 Coverage Approximation Algorithms Example Promotion campaign on a social network We have a social network as a graph. People are more likely to buy a product if they have a friend

More information

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs

Advanced Operations Research Techniques IE316. Quiz 1 Review. Dr. Ted Ralphs Advanced Operations Research Techniques IE316 Quiz 1 Review Dr. Ted Ralphs IE316 Quiz 1 Review 1 Reading for The Quiz Material covered in detail in lecture. 1.1, 1.4, 2.1-2.6, 3.1-3.3, 3.5 Background material

More information

3 No-Wait Job Shops with Variable Processing Times

3 No-Wait Job Shops with Variable Processing Times 3 No-Wait Job Shops with Variable Processing Times In this chapter we assume that, on top of the classical no-wait job shop setting, we are given a set of processing times for each operation. We may select

More information

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes On the Complexity of the Policy Improvement Algorithm for Markov Decision Processes Mary Melekopoglou Anne Condon Computer Sciences Department University of Wisconsin - Madison 0 West Dayton Street Madison,

More information

Distributed minimum spanning tree problem

Distributed minimum spanning tree problem Distributed minimum spanning tree problem Juho-Kustaa Kangas 24th November 2012 Abstract Given a connected weighted undirected graph, the minimum spanning tree problem asks for a spanning subtree with

More information

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 24: Online Algorithms

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 24: Online Algorithms princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 24: Online Algorithms Lecturer: Matt Weinberg Scribe:Matt Weinberg Lecture notes sourced from Avrim Blum s lecture notes here: http://www.cs.cmu.edu/

More information

DESIGN AND ANALYSIS OF ALGORITHMS. Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES

DESIGN AND ANALYSIS OF ALGORITHMS. Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES DESIGN AND ANALYSIS OF ALGORITHMS Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES http://milanvachhani.blogspot.in USE OF LOOPS As we break down algorithm into sub-algorithms, sooner or later we shall

More information

Anarchy is free in network creation

Anarchy is free in network creation Anarchy is free in network creation Ronald Graham Linus Hamilton Ariel Levavi Po-Shen Loh Abstract The Internet has emerged as perhaps the most important network in modern computing, but rather miraculously,

More information

6 Randomized rounding of semidefinite programs

6 Randomized rounding of semidefinite programs 6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can

More information

In this lecture, we ll look at applications of duality to three problems:

In this lecture, we ll look at applications of duality to three problems: Lecture 7 Duality Applications (Part II) In this lecture, we ll look at applications of duality to three problems: 1. Finding maximum spanning trees (MST). We know that Kruskal s algorithm finds this,

More information

FUTURE communication networks are expected to support

FUTURE communication networks are expected to support 1146 IEEE/ACM TRANSACTIONS ON NETWORKING, VOL 13, NO 5, OCTOBER 2005 A Scalable Approach to the Partition of QoS Requirements in Unicast and Multicast Ariel Orda, Senior Member, IEEE, and Alexander Sprintson,

More information

Byzantine Consensus in Directed Graphs

Byzantine Consensus in Directed Graphs Byzantine Consensus in Directed Graphs Lewis Tseng 1,3, and Nitin Vaidya 2,3 1 Department of Computer Science, 2 Department of Electrical and Computer Engineering, and 3 Coordinated Science Laboratory

More information

Lecture Notes: Euclidean Traveling Salesman Problem

Lecture Notes: Euclidean Traveling Salesman Problem IOE 691: Approximation Algorithms Date: 2/6/2017, 2/8/2017 ecture Notes: Euclidean Traveling Salesman Problem Instructor: Viswanath Nagarajan Scribe: Miao Yu 1 Introduction In the Euclidean Traveling Salesman

More information

Optimality certificates for convex minimization and Helly numbers

Optimality certificates for convex minimization and Helly numbers Optimality certificates for convex minimization and Helly numbers Amitabh Basu Michele Conforti Gérard Cornuéjols Robert Weismantel Stefan Weltge May 10, 2017 Abstract We consider the problem of minimizing

More information

Lecture 9: Pipage Rounding Method

Lecture 9: Pipage Rounding Method Recent Advances in Approximation Algorithms Spring 2015 Lecture 9: Pipage Rounding Method Lecturer: Shayan Oveis Gharan April 27th Disclaimer: These notes have not been subjected to the usual scrutiny

More information

11.1 Facility Location

11.1 Facility Location CS787: Advanced Algorithms Scribe: Amanda Burton, Leah Kluegel Lecturer: Shuchi Chawla Topic: Facility Location ctd., Linear Programming Date: October 8, 2007 Today we conclude the discussion of local

More information

Abstract. A graph G is perfect if for every induced subgraph H of G, the chromatic number of H is equal to the size of the largest clique of H.

Abstract. A graph G is perfect if for every induced subgraph H of G, the chromatic number of H is equal to the size of the largest clique of H. Abstract We discuss a class of graphs called perfect graphs. After defining them and getting intuition with a few simple examples (and one less simple example), we present a proof of the Weak Perfect Graph

More information

Coloring 3-Colorable Graphs

Coloring 3-Colorable Graphs Coloring -Colorable Graphs Charles Jin April, 015 1 Introduction Graph coloring in general is an etremely easy-to-understand yet powerful tool. It has wide-ranging applications from register allocation

More information

Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES

Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES DESIGN AND ANALYSIS OF ALGORITHMS Unit 1 Chapter 4 ITERATIVE ALGORITHM DESIGN ISSUES http://milanvachhani.blogspot.in USE OF LOOPS As we break down algorithm into sub-algorithms, sooner or later we shall

More information

Extremal Graph Theory: Turán s Theorem

Extremal Graph Theory: Turán s Theorem Bridgewater State University Virtual Commons - Bridgewater State University Honors Program Theses and Projects Undergraduate Honors Program 5-9-07 Extremal Graph Theory: Turán s Theorem Vincent Vascimini

More information

Combinatorial Gems. Po-Shen Loh. June 2009

Combinatorial Gems. Po-Shen Loh. June 2009 Combinatorial Gems Po-Shen Loh June 2009 Although this lecture does not contain many offical Olympiad problems, the arguments which are used are all common elements of Olympiad problem solving. Some of

More information

12 Introduction to LP-Duality

12 Introduction to LP-Duality 12 Introduction to LP-Duality A large fraction of the theory of approximation algorithms, as we know it today, is built around linear programming (LP). In Section 12.1 we will review some key concepts

More information

1 Better Approximation of the Traveling Salesman

1 Better Approximation of the Traveling Salesman Stanford University CS261: Optimization Handout 4 Luca Trevisan January 13, 2011 Lecture 4 In which we describe a 1.5-approximate algorithm for the Metric TSP, we introduce the Set Cover problem, observe

More information

Clustering. (Part 2)

Clustering. (Part 2) Clustering (Part 2) 1 k-means clustering 2 General Observations on k-means clustering In essence, k-means clustering aims at minimizing cluster variance. It is typically used in Euclidean spaces and works

More information

MAXIMAL PLANAR SUBGRAPHS OF FIXED GIRTH IN RANDOM GRAPHS

MAXIMAL PLANAR SUBGRAPHS OF FIXED GIRTH IN RANDOM GRAPHS MAXIMAL PLANAR SUBGRAPHS OF FIXED GIRTH IN RANDOM GRAPHS MANUEL FERNÁNDEZ, NICHOLAS SIEGER, AND MICHAEL TAIT Abstract. In 99, Bollobás and Frieze showed that the threshold for G n,p to contain a spanning

More information

Applications of the Linear Matroid Parity Algorithm to Approximating Steiner Trees

Applications of the Linear Matroid Parity Algorithm to Approximating Steiner Trees Applications of the Linear Matroid Parity Algorithm to Approximating Steiner Trees Piotr Berman Martin Fürer Alexander Zelikovsky Abstract The Steiner tree problem in unweighted graphs requires to find

More information

15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018

15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018 15-451/651: Design & Analysis of Algorithms October 11, 2018 Lecture #13: Linear Programming I last changed: October 9, 2018 In this lecture, we describe a very general problem called linear programming

More information

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I Instructor: Shaddin Dughmi Announcements Posted solutions to HW1 Today: Combinatorial problems

More information

Joint Entity Resolution

Joint Entity Resolution Joint Entity Resolution Steven Euijong Whang, Hector Garcia-Molina Computer Science Department, Stanford University 353 Serra Mall, Stanford, CA 94305, USA {swhang, hector}@cs.stanford.edu No Institute

More information

Approximation Algorithms: The Primal-Dual Method. My T. Thai

Approximation Algorithms: The Primal-Dual Method. My T. Thai Approximation Algorithms: The Primal-Dual Method My T. Thai 1 Overview of the Primal-Dual Method Consider the following primal program, called P: min st n c j x j j=1 n a ij x j b i j=1 x j 0 Then the

More information