Distributed Submodular Maximization in Massive Datasets. Alina Ene. Joint work with Rafael Barbosa, Huy L. Nguyen, Justin Ward

Similar documents
A survey of submodular functions maximization. Yao Zhang 03/19/2015

Coverage Approximation Algorithms

Optimization of submodular functions

Submodular Optimization

Randomized Composable Core-sets for Distributed Optimization Vahab Mirrokni

Lecture 2. 1 Introduction. 2 The Set Cover Problem. COMPSCI 632: Approximation Algorithms August 30, 2017

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

On Distributed Submodular Maximization with Limited Information

Absorbing Random walks Coverage

Absorbing Random walks Coverage

Submodularity Reading Group. Matroid Polytopes, Polymatroid. M. Pawan Kumar

1 Overview. 2 Applications of submodular maximization. AM 221: Advanced Optimization Spring 2016

A List Heuristic for Vertex Cover

arxiv: v1 [cs.ds] 19 Nov 2018

We ve done. Introduction to the greedy method Activity selection problem How to prove that a greedy algorithm works Fractional Knapsack Huffman coding

Learning Sparse Combinatorial Representations via Two-stage Submodular Maximization

Lecture 9: Pipage Rounding Method

The complement of PATH is in NL

Graphs: Introduction. Ali Shokoufandeh, Department of Computer Science, Drexel University

18 Spanning Tree Algorithms

Approximation Algorithms

Iterative Methods in Combinatorial Optimization. R. Ravi Carnegie Bosch Professor Tepper School of Business Carnegie Mellon University

Greedy Algorithms and Matroids. Andreas Klappenecker

Optimisation While Streaming

Submodular Utility Maximization for Deadline Constrained Data Collection in Sensor Networks

6 Randomized rounding of semidefinite programs

CS 598CSC: Approximation Algorithms Lecture date: March 2, 2011 Instructor: Chandra Chekuri

Greedy Algorithms and Matroids. Andreas Klappenecker

5. Lecture notes on matroid intersection

PTAS for Matroid Matching

Approximation Techniques for Utilitarian Mechanism Design

CS224W: Analysis of Networks Jure Leskovec, Stanford University

Optimization I : Brute force and Greedy strategy

Approximation Guarantees for Adaptive Sampling

Shannon capacity and related problems in Information Theory and Ramsey Theory

A Simple, Greedy Approximation Algorithm for MAX SAT

Outline. CS38 Introduction to Algorithms. Approximation Algorithms. Optimization Problems. Set Cover. Set cover 5/29/2014. coping with intractibility

Greedy Algorithms and Matroids. Andreas Klappenecker

Approximation Algorithms

31.6 Powers of an element

Stanford University CS261: Optimization Handout 1 Luca Trevisan January 4, 2011

Adaptive Caching Algorithms with Optimality Guarantees for NDN Networks. Stratis Ioannidis and Edmund Yeh

Lecture 7: Bipartite Matching

6. Lecture notes on matroid intersection

Algorithm and Complexity of Disjointed Connected Dominating Set Problem on Trees

Submodularity in Speech/NLP

Approximation Algorithms

Theorem 2.9: nearest addition algorithm

Polynomial-Time Approximation Algorithms

Feature Selection. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 262

Viral Marketing and Outbreak Detection. Fang Jin Yao Zhang

Outline. CS38 Introduction to Algorithms. Linear programming 5/21/2014. Linear programming. Lecture 15 May 20, 2014

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi

CSE 202: Design and Analysis of Algorithms Lecture 4

Lecture 9. Semidefinite programming is linear programming where variables are entries in a positive semidefinite matrix.

A Class of Submodular Functions for Document Summarization

Towards Efficient Higher-Order Semidefinite Relaxations for Max-Cut

Greedy algorithms is another useful way for solving optimization problems.

CSC2420 Fall 2012: Algorithm Design, Analysis and Theory

6.842 Randomness and Computation September 25-27, Lecture 6 & 7. Definition 1 Interactive Proof Systems (IPS) [Goldwasser, Micali, Rackoff]

Matt Weinberg. Princeton University

Where have all the cars gone? A model for determining traffic flow throughout a road network

Linear Programming Duality and Algorithms

Rounding procedures in Semidefinite Programming

Approximation Algorithms

Mathematical and Algorithmic Foundations Linear Programming and Matchings

Approximability Results for the p-center Problem

A Lottery Model for Center-type Problems With Outliers

15.082J and 6.855J. Lagrangian Relaxation 2 Algorithms Application to LPs

PCP and Hardness of Approximation

12 Introduction to LP-Duality

Distributed Balanced Clustering via Mapping Coresets

POLYHEDRAL GEOMETRY. Convex functions and sets. Mathematical Programming Niels Lauritzen Recall that a subset C R n is convex if

Towards more efficient infection and fire fighting

Maximum Betweenness Centrality: Approximability and Tractable Cases

Topics in Algorithms 2005 Edge Splitting, Packing Arborescences. Ramesh Hariharan

Solutions for the Exam 6 January 2014

Parallel Algorithms for Geometric Graph Problems Grigory Yaroslavtsev

Models of distributed computing: port numbering and local algorithms

4 Integer Linear Programming (ILP)

Introduction to Algorithms

Important separators and parameterized algorithms

Monotone Closure of Relaxed Constraints in Submodular Optimization: Connections Between Minimization and Maximization: Extended Version

College of Computer & Information Science Fall 2007 Northeastern University 14 September 2007

10. EXTENDING TRACTABILITY

Combinatorial Optimization

Shannon Switching Game

High Dimensional Indexing by Clustering

Submodularity Cuts and Applications

Graph Theory and Optimization Approximation Algorithms

Beyond 1/2-Approximation for Submodular Maximization on Massive Data Streams

Privately Solving Linear Programs

10. EXTENDING TRACTABILITY

Introduction to Mathematical Programming IE406. Lecture 20. Dr. Ted Ralphs

Lecture Notes 2: The Simplex Algorithm

COT 6936: Topics in Algorithms! Giri Narasimhan. ECS 254A / EC 2443; Phone: x3748

Approximation Algorithms: The Primal-Dual Method. My T. Thai

Approximation Algorithms

Graphs and Network Flows IE411. Lecture 13. Dr. Ted Ralphs

Important separators and parameterized algorithms

Transcription:

Distributed Submodular Maximization in Massive Datasets Alina Ene Joint work with Rafael Barbosa, Huy L. Nguyen, Justin Ward

Combinatorial Optimization Given A set of objects V A function f on subsets of V A collection of feasible subsets I Find A feasible subset of I that maximizes f Goal Abstract/general f and I Capture many interesting problems Allow for efficient algorithms

Submodularity We say that a function is submodular if: We say that is monotone if: Alternatively, f is submodular if: for all and Submodularity captures diminishing returns.

Submodularity Examples of submodular functions: The number of elements covered by a collection of sets Entropy of a set of random variables The capacity of a cut in a directed or undirected graph Rank of a set of columns of a matrix Matroid rank functions Log determinant of a submatrix

Example: Multimode Sensor Coverage We have distinct locations where we can place sensors Each sensor can operate in different modes, each with a distinct coverage profile Find sensor locations, each with a single mode to maximize coverage

Example: Identifying Representatives In Massive Data

Example: Identifying Representative Images We are given a huge set X of images. Each image is stored multidimensional vector. We have a function d giving the difference between two images. We want to pick a set S of at most k images to minimize the loss function: Suppose we choose a distinguished vector e 0 (e.g. 0 vector), and set: The function f is submodular. Our problem is then equivalent to maximizing f under a single cardinality constraint.

Need for Parallelization Datasets grow very large TinyImages has 80M images Kosarak has 990K sets Need multiple machines to fit the dataset Use parallel frameworks such as MapReduce

Problem Definition Given set V and submodular function f Hereditary constraint I (cardinality at most k, matroid constraint of rank k, ) Find a subset that satisfies I and maximizes f Parameters n = V k : max size of feasible solutions m : number of machines

Greedy Algorithm Initialize S = {} While there is some element x that can be added to S: Add to S the element x that maximizes the marginal gain Return S

Greedy Algorithm Approximation Guarantee: 1-1/e for a cardinality constraint 1/2 for a matroid constraint Runtime: O(nk) Need to recompute marginals each time an element is added Not good for large data sets

Distributed Greedy Mirzasoleiman, Karbasi, Sarkar, Krause '13

Mirzasoleiman, Karbasi, Sarkar, Krause '13 Performance of Distributed Greedy Only requires 2 rounds of communication Approximation ratio is: (where m is number of machines) If we use the optimal algorithm on each machine in both phases, we can still only get:

Performance of Distributed Greedy If we use the optimal algorithm on each machine in both phases, we can still only get: In fact, we can show that using greedy gives: Why? The problem doesn't have optimal substructure. Better to run greedy in round 1 instead of the optimal algorithm.

Revisiting the Analysis Can construct bad examples for Greedy/optimal Lower bound for any poly(k) coresets (Indyk et al. 14) Yet the distributed greedy algorithm works very well on real instances Why?

Power of Randomness Randomized distributed Greedy Distribute the elements of V randomly in round 1 Select the best solution found in rounds 1 & 2 Theorem: If Greedy achieves a C approximation, randomized distributed Greedy achieves a C/2 approximation in expectation.

Intuition If elements in OPT are selected in round 1 with high probability Most of OPT is present in round 2 so solution in round 2 is good If elements in OPT are selected in round 1 with low probability OPT is not very different from typical solution so solution in round 1 is good

Analysis (Preliminaries) Greedy Property: Suppose: x is not selected by greedy on S {x} y is not selected by greedy on S {y} Then: x and y are not selected by greedy on S {x,y} Lovasz extension : convex function on [0,1] V that agrees with on integral vectors.

Analysis (Sketch) Let X be a random 1/m sample of V For e in OPT, let p e be the probability (over choice of X) that e is selected by Greedy on X {e} Then, expected value of elements of OPT on the final machine is On the other hand, expected value of rejected elements is

Analysis (Sketch) The final greedy solution T satisfies: The best single machine solution S satisfies: Altogether, we get an approximation in expectation of:

Generality What do we need for the proof? Monotonicity and submodularity of f Heredity of constraint Greedy property The result holds in general any time greedy is an -approximation for a hereditary, constrained submodular maximization problem.

Non-monotone Functions In the first round, use Greedy on each machine In the second round, use any algorithm on the last machine We still obtain a constant factor approximation for most problems

Tiny Image Experiments (n = 1M, m = 100)

Matroid Coverage Experiments Matroid Coverage (n=900, r=5) Matroid Coverage (n=100, r=100) It's better to distribute ellipses from each location across several machines!

Future Directions Can we relax the greedy property further? What about non-greedy algorithms? Can we speed up the final round, or reduce the number machines required? Better approximation guarantees?