Spectral Graph Sparsification: overview of theory and practical methods. Yiannis Koutis. University of Puerto Rico - Rio Piedras

Similar documents
SDD Solvers: Bridging theory and practice

Lecture 25: Element-wise Sampling of Graphs and Linear Equation Solving, Cont. 25 Element-wise Sampling of Graphs and Linear Equation Solving,

SPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING

Laplacian Paradigm 2.0

Lesson 2 7 Graph Partitioning

Parallel Graph Algorithms. Richard Peng Georgia Tech

Combinatorial meets Convex Leveraging combinatorial structure to obtain faster algorithms, provably

Lecture 11: Clustering and the Spectral Partitioning Algorithm A note on randomized algorithm, Unbiased estimates

Johnson-Lindenstrauss Lemma, Random Projection, and Applications

Homomorphic Sketches Shrinking Big Data without Sacrificing Structure. Andrew McGregor University of Massachusetts

Graphs in Machine Learning

CS7540 Spectral Algorithms, Spring 2017 Lecture #2. Matrix Tree Theorem. Presenter: Richard Peng Jan 12, 2017

Algebraic Graph Theory- Adjacency Matrix and Spectrum

Solvers for Linear Systems in Graph Laplacians

Lecture 27: Fast Laplacian Solvers

Clustering in Networks

Exponential Start Time Clustering and its Applications in Spectral Graph Theory. Shen Chen Xu

Lecture 5: Graphs. Rajat Mittal. IIT Kanpur

SAMG: Sparsified Graph-Theoretic Algebraic Multigrid for Solving Large Symmetric Diagonally Dominant (SDD) Matrices

Social-Network Graphs

Spectral Clustering X I AO ZE N G + E L HA M TA BA S SI CS E CL A S S P R ESENTATION MA RCH 1 6,

REGULAR GRAPHS OF GIVEN GIRTH. Contents

The Revolution in Graph Theoretic Optimization Problem

Preconditioning Linear Systems Arising from Graph Laplacians of Complex Networks

CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning

Towards Optimal Dynamic Graph Compression

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1

Parallel Algorithms for (PRAM) Computers & Some Parallel Algorithms. Reference : Horowitz, Sahni and Rajasekaran, Computer Algorithms

BACKGROUND: A BRIEF INTRODUCTION TO GRAPH THEORY

The Matrix-Tree Theorem and Its Applications to Complete and Complete Bipartite Graphs

Integrated Math I. IM1.1.3 Understand and use the distributive, associative, and commutative properties.

Mining Social Network Graphs

Efficient and practical tree preconditioning for solving Laplacian systems

Spectral Sparsification in Spectral Clustering

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014

Graph drawing in spectral layout

UNIVERSITY OF PUERTO RICO RIO PIEDRAS CAMPUS COLLEGE OF NATURAL SCIENCE DEPARTMENT OF MATHEMATICS

Definition Example Solution

Parallel Greedy Matching Algorithms

Speeding up MWU based Approximation Schemes and Some Applications

An Introduction to Graph Theory

Applying the weighted barycentre method to interactive graph visualization

Big Data Management and NoSQL Databases

Kirchhoff's theorem. Contents. An example using the matrix-tree theorem

My favorite application using eigenvalues: partitioning and community detection in social networks

Big Data Analytics. Special Topics for Computer Science CSE CSE Feb 11

Community Detection. Community

Chapter 2. Spectral Graph Drawing

Depth First Search A B C D E F G A B C 5 D E F 3 2 G 2 3

Math 778S Spectral Graph Theory Handout #2: Basic graph theory

CHAPTER 2. Graphs. 1. Introduction to Graphs and Graph Isomorphism

A Course in Machine Learning

What is a minimal spanning tree (MST) and how to find one

Algorithms design under a geometric lens Spring 2014, CSE, OSU Lecture 1: Introduction

On Covering a Graph Optimally with Induced Subgraphs

Cospectral graphs. Presented by: Steve Butler. 21 April 2011

Introduction to Machine Learning

Visual Representations for Machine Learning

Graph Sparsification, Spectral Sketches, and Faster Resistance Computation, via Short Cycle Decompositions

Kernighan/Lin - Preliminary Definitions. Comments on Kernighan/Lin Algorithm. Partitioning Without Nodal Coordinates Kernighan/Lin

Trees. 3. (Minimally Connected) G is connected and deleting any of its edges gives rise to a disconnected graph.

TELCOM2125: Network Science and Analysis

CS 561, Lecture 9. Jared Saia University of New Mexico

The 4/3 Additive Spanner Exponent is Tight

Sublinear Algorithms Lectures 1 and 2. Sofya Raskhodnikova Penn State University

Solving problems on graph algorithms

Separators in High-Genus Near-Planar Graphs

On the Rigidity of Ramanujan Graphs

CONNECTIVITY CHECK IN 3-CONNECTED PLANAR GRAPHS WITH OBSTACLES

Elementary Graph Algorithms. Ref: Chapter 22 of the text by Cormen et al. Representing a graph:

Latest on Linear Sketches for Large Graphs: Lots of Problems, Little Space, and Loads of Handwaving. Andrew McGregor University of Massachusetts

CSCI5070 Advanced Topics in Social Computing

DEGENERACY AND THE FUNDAMENTAL THEOREM

The number of spanning trees of plane graphs with reflective symmetry

Lecture 19: Graph Partitioning

6 Randomized rounding of semidefinite programs

Social Network Analysis

arxiv: v1 [cs.ds] 1 Jan 2015

(Sparse) Linear Solvers

Spectral Clustering on Handwritten Digits Database

Graph Partitioning for High-Performance Scientific Simulations. Advanced Topics Spring 2008 Prof. Robert van Engelen

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1

Vertex Magic Total Labelings of Complete Graphs 1

Efficient and practical tree preconditioning for solving Laplacian systems

Cpt S 223 Course Overview. Cpt S 223, Fall 2007 Copyright: Washington State University

Math 485, Graph Theory: Homework #3

Vertex Magic Total Labelings of Complete Graphs

CMPSCI 711: More Advanced Algorithms

12/5/17. trees. CS 220: Discrete Structures and their Applications. Trees Chapter 11 in zybooks. rooted trees. rooted trees

Notes for Lecture 20

Sparse Matrices and Graphs: There and Back Again

CSE 547: Machine Learning for Big Data Spring Problem Set 2. Please read the homework submission policies.

CS 5220: Parallel Graph Algorithms. David Bindel

CS 310 Advanced Data Structures and Algorithms

Spectral Compression of Mesh Geometry

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Geometric transformations assign a point to a point, so it is a point valued function of points. Geometric transformation may destroy the equation

Notes for Lecture 24

CS388C: Combinatorics and Graph Theory

AMS526: Numerical Analysis I (Numerical Linear Algebra)

Transcription:

Spectral Graph Sparsification: overview of theory and practical methods Yiannis Koutis University of Puerto Rico - Rio Piedras

Graph Sparsification or Sketching Compute a smaller graph that preserves some crucial property of the input Motivation: Computational efficiency with approximation guarantees BFS: Breadth First Spanning Tree Spanner: Spanning subgraph that approximately preserves distances

Spectral Graph Sparsification Compute a smaller graph that preserves some crucial property of the input We want to approximately preserve the eigenvalues and eigenvectors of the graph Laplacian Motivation: Speed-up many clustering and partitioning algorithms based on computing Laplacian eigenvectors

Spectral Graph Sparsification Compute a smaller graph that preserves some crucial property of the input We want to approximately preserve the quadratic form x T Lx of the Laplacian L Implies spectral approximations for both the Laplacian and the normalized Laplacian

The Graph Laplacian 30 2 1 1 20 15

Spectral Sparsification by Picture H is a reweighted subgraph of G H is obtained using randomness (sampling)

Combinatorial sketching Spectral sketching Linear system solvers Better combinatorial sketching

Outline Sums of random positive matrices Combinatorial sketching to incremental sparsification Incremental sparsification for solving Parallel and distributed sparsification Deep sparsification by effective resistances A heuristic for better clustering

Matrix Ordering Whenever for all vectors x we have We write G ¹ H In this notation, a spectral sparsifier H satisfies

Sums of random matrices [Tropp 12, adapted]: 1. Let S be a nxn PSD matrix and Y 1,, Y k be independent random PSD matrices also of size nxn. 2. Let Y = S+ i Y i 3. Let Z = E[Y] 4. Suppose Y i ¹ R Z S : Combinatorial Sketch Y i : Edges R: should be O(C/log n)

Outline Sums of random positive matrices Combinatorial sketching to incremental sparsification Incremental sparsification for solving Parallel and distributed sparsification Deep sparsification by effective resistances A heuristic for better clustering

A simple algorithm 1. Compute a spanner S : 2. Let H:= S 3. For every edge e not in H: H:=H + k*e, with probability 1/k H has O(nlog n) + m/k edges n: number of vertices m: number of edges

and a simple proof Suppose the graph G has n vertices and m edges For simplicity we let G be unweighted (generalization is easy) By definition of spanner, for every edge e of G, there is a path p e in S that joins the two endpoints of e and has length logn. Algebraically, if G e is the Laplacian of edge e:

and a simple proof Apply Tropp s Theorem on: Combinatorial Sketch: Samples:

Incremental Sparsification for Solving H has O(nlog n) + m/k edges [Spielman and Teng] If we can construct H with same guarantees but only n+m/k edges then we can solve linear systems on Laplacians in O(mlog 2 n) time [K, Miller Peng 10] Use low-stretch tree instead of spanner. It preserves distances on average. Use sampling with skewed probability distribution.

Incremental Sparsification for Solving [K, Miller Peng 10,11] If A is a symmetric diagonally dominant matrix then an approximate solution to Ax= b can be computed in O(m log n log log n log (1/²)) time, where ² is the required precision Fact: Approximations to the j first eigenvectors can be computed via solving O(j log n) linear systems

Outline Sums of random positive matrices Combinatorial sketching to incremental sparsification Incremental sparsification for solving Parallel and distributed sparsification Deep sparsification by effective resistances A heuristic for better clustering

Parallel and Distributed Sparsification 1. Can we do better than incremental sparsification? 2. Is there a parallel sparsification algorithm? 3. Is there a distributed sparsification algorithm? Spanners hold the key to questions #2, #3 There are very efficient parallel and distributed algorithms for computing spanners. From this we get parallel and distributed incremental sparsification. This doesn t itself imply parallel solvers. The main problem is question #1.

Parallel and Distributed Sparsification A better combinatorial sketch: t-bundle spanner: A collection of graphs S 1.S t such that S i is a spanner for G (S 1 +.+ S i-1 ) A t-bundle spanner can be computed with t sequential calls to a spanner computation algorithm. If t is small then the algorithm remains efficient (polylogarithmic parallel time and distributed rounds)

A simple algorithm 1. Compute O(log 4 n)-bundle spanner S 2. Let H:= S 3. For every edge e not in H: H:=H + 2*e, with probability 1/2 H has O(nlog 5 n) + m/2 edges n: number of vertices m: number of edges

A simple algorithm H has O(nlog 5 n) + m/2 edges Small size reduction (factor of 2) Very tight spectral approximation Repeat recursively on H. In O(log n) rounds we get O(nlog 6 n) edges and a constant spectral approximation.

A parallel solver This is the best known parallel sparsification routine. Improves the total work guarantees of a parallel solver recently described by Peng and Spielman. Parallel solver that works in polylogarithmic time and does O(mlog 3 n) work.

Outline Sums of random positive matrices Combinatorial sketching to incremental sparsification Incremental sparsification for solving Parallel and distributed sparsification Deep sparsification by effective resistances A heuristic for better clustering

Spielman-Srivastava: Deep sparsification Spielman and Srivastava proved: There is a graph H with edges such that The algorithm is based on sampling edges with probabilities proportional to the effective resistances of the edges in the graph Proof uses similarly Tropp s theorem

Combinatorial sketching Spectral sketching Linear system solvers Better combinatorial sketching

The Incidence Matrix 30 2 1 1 20 15 W is the diagonal matrix containing the square roots of edge weights

Spielman-Srivastava: Deep sparsification Effective resistances can be approximated closely by solving O(log n) linear Laplacian systems as follows: 1. Let L be the Laplacian of the graph 2. Let B be the incident matrix of the graph 3. Let Q be a random Johnson-Lindenstrauss projection of size m x O(log n) 4. Solve the systems L X = B T Q 5. Effective resistance between vertices i and j is equal to the X i X i 2 where X i is the ith row of X The solution X is a n x O(log n) matrix. Each row can be interpreted as an embedding of corresponding vertex to the O(log n)-dimensional Euclidean space. Let s call this the effective resistance embedding.

http://ccom.uprrp.edu/~ikoutis/spectralalgorithms.htm

Heuristic: clustering based on the effective resistance embedding Small effective resistance for an edge e means that there are a lot of short connections between the two endpoints. Points that are close in the geometric embedding should be close in this connectivity sense in the graph. Idea: Produce a k-clustering of the graph by running k-means on the effective resistance embedding Produced clusterings appear to have better properties than clusterings based on geometric embeddings using the k-first eigenvectors. It is also much faster for most values of k.

Visualization of unweighted social network graph

Visualization of the same graph using effective resistances as weights

Combinatorial sketching Spectral sketching Linear system solvers Better combinatorial sketching

Spectral Sparsification vs Algebraic Sketching The sparsifier H of G has the form Here S is a diagonal matrix containing the reweight factors of the edges and B is the incidence matrix for G Algebraic sketching : instead of solving a regression problem with a matrix B, solve instead one with SB where S is a sparse projection matrix Goal is

Spectral Sparsification vs Algebraic Sketching So, spectral graph sparsification is a special instance of algebraic sketching Algebraic sketching takes a very tall and thin matrix and finds a nearly equivalent tall and thin matrix In the graph sparsification case graph combinatorics allow for a much tighter control on size reduction