Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees

Similar documents
Probabilistic embedding into trees: definitions and applications. Fall 2011 Lecture 4

Metric Techniques and Approximation Algorithms. Anupam Gupta Carnegie Mellon University

A tight bound on approximating arbitrary metrics by tree metric Reference :[FRT2004]

Embedding Bounded Bandwidth Graphs into l p

Embedding Bounded Bandwidth Graphs into l 1

THREE LECTURES ON BASIC TOPOLOGY. 1. Basic notions.

Sublinear Time and Space Algorithms 2016B Lecture 7 Sublinear-Time Algorithms for Sparse Graphs

Improved Embeddings of Graph Metrics into Random Trees

Approximation Algorithms

Embedding Bounded Bandwidth Graphs into l 1

Randomized Optimization Problems on Hierarchically Separated Trees

Lecture 6: Linear Programming for Sparsest Cut

CSE 548: Analysis of Algorithms. Lecture 13 ( Approximation Algorithms )

Randomly Removing g Handles at Once

Lecture 8: The Traveling Salesman Problem

CPSC 536N: Randomized Algorithms Term 2. Lecture 10

arxiv: v4 [cs.ds] 25 May 2017

8 Matroid Intersection

Fixed-Parameter Algorithms, IA166

1 The Traveling Salesperson Problem (TSP)

Topic: Local Search: Max-Cut, Facility Location Date: 2/13/2007

Solutions for the Exam 6 January 2014

The External Network Problem

val(y, I) α (9.0.2) α (9.0.3)

Exact Algorithms Lecture 7: FPT Hardness and the ETH

6 Randomized rounding of semidefinite programs

3 No-Wait Job Shops with Variable Processing Times

Jessica Su (some parts copied from CLRS / last quarter s notes)

CMSC 451: Lecture 22 Approximation Algorithms: Vertex Cover and TSP Tuesday, Dec 5, 2017

Discrete Mathematics Course Review 3

1 Minimum Cut Problem

Treewidth and graph minors

Chapter 24. Shortest path problems. Chapter 24. Shortest path problems. 24. Various shortest path problems. Chapter 24. Shortest path problems

4. Definition: topological space, open set, topology, trivial topology, discrete topology.

Points covered an odd number of times by translates

CS270 Combinatorial Algorithms & Data Structures Spring Lecture 19:

THE GROWTH OF LIMITS OF VERTEX REPLACEMENT RULES

Bandwidth Approximation of Many-Caterpillars

The k-center problem Approximation Algorithms 2009 Petros Potikas

Vertex Cover Approximations

Lecture 2 The k-means clustering problem

Fast Clustering using MapReduce

A SIMPLE APPROXIMATION ALGORITHM FOR NONOVERLAPPING LOCAL ALIGNMENTS (WEIGHTED INDEPENDENT SETS OF AXIS PARALLEL RECTANGLES)

Online Algorithms. Lecture 11

Notes on Binary Dumbbell Trees

MATH 54 - LECTURE 10

Simplicial Hyperbolic Surfaces

Traveling Salesman Problem (TSP) Input: undirected graph G=(V,E), c: E R + Goal: find a tour (Hamiltonian cycle) of minimum cost

Theorem 2.9: nearest addition algorithm

On the Max Coloring Problem

1 Variations of the Traveling Salesman Problem

Improved algorithms for constructing fault-tolerant spanners

11.1 Facility Location

9 Bounds for the Knapsack Problem (March 6)

Treaps. 1 Binary Search Trees (BSTs) CSE341T/CSE549T 11/05/2014. Lecture 19

Lecture Overview. 2 Shortest s t path. 2.1 The LP. 2.2 The Algorithm. COMPSCI 530: Design and Analysis of Algorithms 11/14/2013

Copyright 2000, Kevin Wayne 1

Lecture 5: Duality Theory

On the Complexity of the Policy Improvement Algorithm. for Markov Decision Processes

Online algorithms for clustering problems

DO NOT RE-DISTRIBUTE THIS SOLUTION FILE

Non-context-Free Languages. CS215, Lecture 5 c

Definition For vertices u, v V (G), the distance from u to v, denoted d(u, v), in G is the length of a shortest u, v-path. 1

Sources for this lecture. 3. Matching in bipartite and general graphs. Symmetric difference

Math 734 Aug 22, Differential Geometry Fall 2002, USC

CS369G: Algorithmic Techniques for Big Data Spring

Approximation Algorithms: The Primal-Dual Method. My T. Thai

CS599: Convex and Combinatorial Optimization Fall 2013 Lecture 14: Combinatorial Problems as Linear Programs I. Instructor: Shaddin Dughmi

Lecture 2 - Introduction to Polytopes

On Minimum Sum of Radii and Diameters Clustering

22 Elementary Graph Algorithms. There are two standard ways to represent a

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1

Maintaining Approximate Minimum Steiner Tree and k-center for Mobile Agents in a Sensor Network

Hamiltonian cycles in bipartite quadrangulations on the torus

Point-Set Topology 1. TOPOLOGICAL SPACES AND CONTINUOUS FUNCTIONS

Approximate Min-Max Theorems for Steiner Rooted-Orientations of Graphs and Hypergraphs

Problem Set 1. Solution. CS4234: Optimization Algorithms. Solution Sketches

Algorithms for Grid Graphs in the MapReduce Model

31.6 Powers of an element

Technische Universität München, Zentrum Mathematik Lehrstuhl für Angewandte Geometrie und Diskrete Mathematik. Combinatorial Optimization (MA 4502)

CS261: A Second Course in Algorithms Lecture #16: The Traveling Salesman Problem

Testing Continuous Distributions. Artur Czumaj. DIMAP (Centre for Discrete Maths and it Applications) & Department of Computer Science

Partitions and Packings of Complete Geometric Graphs with Plane Spanning Double Stars and Paths

The Encoding Complexity of Network Coding

Monotone Paths in Geometric Triangulations

V1.0: Seth Gilbert, V1.1: Steven Halim August 30, Abstract. d(e), and we assume that the distance function is non-negative (i.e., d(x, y) 0).

Upper bounds and algorithms for parallel knock-out numbers

Solution for Homework set 3

Algorithms for Bioinformatics

We show that the composite function h, h(x) = g(f(x)) is a reduction h: A m C.

Algorithms design under a geometric lens Spring 2014, CSE, OSU Lecture 1: Introduction

Sources for this lecture 2. Shortest paths and minimum spanning trees

Numerical Optimization

Local Search Approximation Algorithms for the Complement of the Min-k-Cut Problems

In this lecture, we ll look at applications of duality to three problems:

Online Facility Location

22 Elementary Graph Algorithms. There are two standard ways to represent a

Computer Science 280 Fall 2002 Homework 10 Solutions

Approximating Node-Weighted Multicast Trees in Wireless Ad-Hoc Networks

1 (15 points) LexicoSort

Transcription:

Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees Lior Kamma 1 Introduction Embeddings and Distortion An embedding of a metric space (X, d X ) into a metric space (Y, d Y ) is a map f : X Y Its (bi-lipschitz) distortion is the least D 1 such that x, y X d X (x, y) d Y (f(x), f(y)) D d X (x, y) Some related results previously seen in class Claim Every n-point metric space embeds isometrically (ie, with distortion 1) into l n Theorem (Bourgain 1985) Every n-point metric space embeds into l 2 O(log n) with distortion Theorem (Johnson-Lindenstrauss) Every n-point metric subspace of l d 2 embeds into l k 2 with distortion (1 + ε), where k = O(ε 2 log n) Tree Metrics {w e } e E Consider an undirected graph G = (V, E) with non-negative edge weights Exercise: Show that the function d G : V V R, which maps every pair x, y V to the length of a shortest path between x and y in G wrt w, is a metric on V A metric space (Y, d Y ) is called a tree metric space if there exists a tree G such that Y embeds isometrically into G Dream Goal : Embed an arbitrary metric space (X, d X ) into a tree metric space with small distortion We first note that every finite tree metric space can be embedded isomet- Motivation rically into l 1 Exercise: Prove it Additionally, many optimization and online problems involve a metric defined on a set of points It is often useful to embed a metric space into a simpler one while keeping 1 / 7

the distances approximately Specifically, many such problems can be efficiently solved or better approximated on trees Bear the following example, called k-median, in mind You are given a metric space (X, d X ) and an integer k The goal is to choose a set S X of size at most k, that minimizes the objective function x X d X(x, S) This problem is known to be NP-Hard, however it can be solved optimally on trees in polynomial time The heuristic is as follows Embed X into a tree metric Y, solve the problem on Y, and construct a respective solution in X Details are omitted at this point, mainly due to the fact that, unfortunately, this approach does not work so well Embedding a Cycle into a Single Tree Let (C n, d Cn ) denote the shortest-path metric on an unweighted n-cycle One can easily show show that embedding the cycle into a spanning tree incurs a distortion D Ω(n) In fact, Rabinovich and Raz [RR98] showed that every embedding of the cycle into a tree (not necessarily a spanning tree, and may have additional vertices) incurs distortion Ω(n) 2 Randomized Embeddings However, not all is lost If we consider a random embedding of C n, then we can bound the distortion in expectation Let T be the random tree that results from deleting a single edge of C n chosen uniformly at random Notice that this embedding satisfies the following two properties (proved in class) 1 For every x, y C n d Cn (x, y) d T (x, y) 2 For every x, y C n E[d T (x, y)] 2d Cn (x, y) Exercise: Extend the result to a weighted cycle New Goal Embed an arbitrary metric space (X, d X ) into a random dominating tree metric with small expected distortion In fact, we will show a somewhat stronger result Definition 1 A k-hierarchically well-separated tree (k-hst) is a rooted weighted tree T = (V (T ), E(T )) satisfying the following properties 1 For every node v V (T ), all edges connecting v to a child are of equal weight 2 The edge weight along a path from the root to a leaf decrease by a factor of at least k 2 / 7

Theorem 1 ([FRT04]) Let (X, d X ) be an n-point metric space There exists a randomized polynomial-time algorithm that embeds X into the set of leaves of a 2-HST T = (V (T ), E(T )) such that the following holds (we may assume that X V (T )) 1 For every x, y X d(x, y) d T (x, y) 2 For every x, y X E[d T (x, y)] O(log n)d X (x, y) Note that since the distortion is bounded in expectation, we can still apply the approximation heuristic considered earlier for problems in which the objective function is linear Back to k-median Lemma 1 The k-median problem can be solved efficiently on the metric space induced by the set of leaves of a 2-HST Exercise: Prove Lemma 1 Hint: Use dynamic programming Corollary 1 There exists a randomized approximation algorithm for the k-median problem with expected ratio O(log n) proof sketch Given a metric space (X, d X ) and an integer k, we apply Theorem 1 and randomly embed X into a 2-HST T We solve the problem on the leaves of T and return the solution 3 Partitions, Laminar Families and Trees Definition 2 A set-family L 2 X is called laminar if for every A, B L, if A B then A B or B A A laminar family L 2 X such that {x} L for all x X, induces a tree T such that V (T ) = L and the leaves of T are exactly {{x} : x X} in a straightforward manner We can construct a laminar family by repeatedly partitioning X In order to make sure the algorithm halts, we can, eg decrease the diameter of the sets in the partition in each iteration Let Π be a partition of X Every A Π is called a cluster, and for every x X, let Π(x) denote the unique cluster A Π such that x S Denote the diameter of X by By scaling we may assume without loss of generality that min x,y X d X (x, y) = 1 3 / 7

Input: X Output: A laminar family L 2 X such that {{x} : x X} L 1: Π 0 {X}, L {X} 2: for i = 1 to log do 3: Π i 4: for all A Π i 1 do 5: if A > 1 then 6: Let Π be a partition of A into clusters of diameter at most 2 i 7: Π i Π i Π 8: L L Π 9: return L Algorithm 1: Constructing a Laminar Family It remains to show how to construct the partitions Π i, i [log ], and how to set the weights of the tree edges 4 From Low-Diameter Decompositions to Low-Distortion Embeddings Definition 3 A metric space (X, d X ) is called β-decomposable for β > 0 if for every > 0 there is a probability distribution µ over partitions of X, satisfying the following properties (a) Diameter Bound: For every Π supp(µ) and A Π, diam(a) (b) Separation: For every x, y X, Pr [Π(x) Π(y) ] β dx(x, y) Π µ Theorem 2 ([Bar96], [FRT04]) Every n-point metric space is 8 log n-decomposable In fact, Fakcharoenphol, Rao and Talwar [FRT04] gave a somewhat stronger result, which will prove essential in the analysis of the embedding We replace the separation property in Definition 3 by the following, stronger requirement (b ) For every x, y X, if d X (x, y) < 8 then Pr [Π(x) Π(y) ] d X(x, y) B({x, y}, /2) 8 log Π µ B({x, y}, /8), where B({x, y}, r) = {z X : d X ({x, y}, z) r} for all r > 0 4 / 7

We can now update Algorithm 1 and construct the tree embedding Input: X Output: A 2-HST with X being the set of leaves 1: Π 0 {X} 2: V (T ) Π 0, E(T ) 3: for i = 1 to log do 4: Π i 5: for all A Π i 1 do 6: if A > 1 then 7: Let Π be a random partition of A as in Theorem 2 with = 2 i 8: Π i Π i Π 9: V (T ) V (T ) Π 10: Add to E(T ) an edge from A to every cluster in Π, of weight 11: return T Algorithm 2: Constructing a Random Embedding into a 2-HST Applying Theorem 2, we now turn to prove Theorem 1 Note that the leaves of T are exactly the sets {x} for all x X, and thus for every x X, we can identify {x} V (T ) with x Clearly T is a 2-HST Consider next x, y X and let i 0 be the unique integer such that d X (x, y) (2 i 0, 2 (i 0 1) ], and let i be the first index for which Π i (x) Π i (y) By the diameter bound of the partition we get that i i 0 We therefore conclude the following Claim 1 d T (x, y) d X (x, y) Proof d T (x, y) 2 2 i 2 i 0+1 d X (x, y) Claim 2 d T (x, y) 2 i +2 Proof Denote by u V (T ) the least common ancestor of x, y Consider the path from u to x Since T is a 2-HST we get that the length of the path is at most i=i 2 i = 2 i +1 The length of the xy-path in T is at most twice as long The following claim concludes the proof of Theorem 1 Corollary 2 E[d T (x, y)] O(log n)d X (x, y) Proof Since 1 i i 0, then E[d T (x, y)] = i 0 i=1 E[d T (x, y) i = i] Pr[i = i] By Claim 2, E[d T (x, y) i = i] 2 i+2, and by Theorem 2, Pr[i = i] d X(x,y) Therefore E[d T (x, y)] i 0 i=1 2 i+2 dx(x, y) 2 i log B({x, y}, 2 i 1 ) B({x, y}, 2 i 3 ) = 4d X(x, y) 5 / 7 log B({x,y},2 i /2) 2 i B({x,y},2 i /8) i 0 i=1 log B({x, y}, 2 i 1 ) B({x, y}, 2 i 3 )

All but a constant number of elements of the sum are canceled, and therefore E[d T (x, y)] O(log n)d X (x, y) 5 Randomized Low-Diameter Decompositions We now turn to prove Theorem 2 The following algorithm samples a partition of X We will show that the distribution induced by the algorithm satisfies the conditions of the theorem Input: X, Output: A partition Π as in Theorem 2 1: Π 2: let π be a random ordering of X 3: independently choose R (/4, /2] uniformly at random 4: for all j [n] do 5: let B j = B(π(j), R) 6: let C j = B j \ j <j B j 7: if C j then Π Π {C j } 8: return Π Algorithm 3: Constructing a Random Partition Clearly for every C Π, diam(c) Fix x, y X, and let x 1, x 2,, x n be an ordering of X in ascending distance from {x, y} (breaking ties arbitrarily) Fix j [n] We say that x j settles x, y if B(x j, R) is the first ball (in the order induced by π) that has non-empty intersection with {x, y} We say that x j cuts x, y if B(x j, R) {x, y} = 1, and x j separates x, y if x j both settles x, y and cuts x, y Notice that the event that x j cuts x, y depends only on the choice of R and is independent of the choice of π Assume, without loss of generality that d X (j, x) d X (j, y) Claim 3 Pr[x j separates x, y] 1 j 4d X(x,y) Proof First note that Pr[x j separates x, y] = Pr[x j separates x, y x j cuts x, y] Pr[x j cuts x, y] Note that x j cuts x, y if and only if R [d X (j, x), d X (j, y)) Since R is uniformly distributed over (/4, /2], and from the triangle inequality we get that Pr[x j cuts x, y] = Pr[R [d X (j, x), d X (j, y)) ] d X(j, y) d X (j, x) /4 4d X(x, y) Conditioned on x j cutting x, y, assume toward contradiction that there exists j < j such that x j precedes x j in the order induced by π Since d X (x j, {x, y}) d X (x j, {x, y}) = 6 / 7

d X (x j, x) R, it follows that {x, y} B(x j, R) and therefore x j does not settle x, y, a contradiction Therefore, Pr[x j separates x, y x j cuts x, y] Pr[x j precedes x j for all j < j] 1 j Since Pr[Π(x) Π(y)] j [n] Pr[x j separates x, y] we get that Pr[Π(x) Π(y)] j [n] 4d X (x, y) j 4 d X(x, y) (log n + 1) d X(x, y) 8 log n In order to get a stronger result, we need a more delicate analysis Assume that d X (x, y) /8, then if x j B({x, y}, /8), then Pr[x j separates x, y] = 0 In addition, if x j / B({x, y}, /2), then Pr[x j separates x, y] = 0 Therefore Pr[Π(x) Π(y)] j B({x,y},/2)\B({x,y},/8) 4d X (x, y) j d X(x, y) 4 log B({x, y}, /2) B({x, y}, /8) Exercise: A metric space (X, d X ) is called β-padded-decomposable for β > 0 if for every > 0 there is a probability distribution µ over partitions of X, satisfying the following properties (a) Diameter Bound: For every Π supp(µ) and A Π, diam(a) (b) Padding: For every x X and ε < /8, ε Pr [B(x, ε) Π(x)] β Π µ Show that every n-point metric space is O(log n)-padded-decomposable References [Bar96] Y Bartal Probabilistic approximation of metric spaces and its algorithmic applications In 37th Annual Symposium on Foundations of Computer Science, pages 184 193 IEEE, 1996 [FRT04] J Fakcharoenphol, S Rao, and K Talwar A tight bound on approximating arbitrary metrics by tree metrics J Comput Syst Sci, 69(3):485 497, 2004 [RR98] Y Rabinovich and R Raz Lower bounds on the distortion of embedding finite metric spaces in graphs Discrete Comput Geom, 19(1):79 94, 1998 7 / 7