Spectral Clustering and Community Detection in Labeled Graphs

Size: px
Start display at page:

Download "Spectral Clustering and Community Detection in Labeled Graphs"

Transcription

1 Spectral Clustering and Community Detection in Labeled Graphs Brandon Fain, Stavros Sintos, Nisarg Raval Machine Learning (CompSci 571D / STA 561D) December 7, 2015 {btfain, nisarg, ssintos} at cs.duke.edu Abstract We study spectral clustering techniques to learn community structures in labeled random graphs where edge labels from a label set L = {1,..., L} are drawn according to discrete probability distributions parametrized by community membership of the two end-nodes of the edge. This is a strict generalization of the standard stochastic block model for community detection. 1 Introduction and Related Work Graph partitioning is a fundamental problem in computer science first posed as the minimum cut problem: given a graph G = (V, E), one tries to find a partitioning of the vertices into two clusters S 1, S 2 such that S 1 S 2 = V and S 1 S 2 = so as to minimize the number of edges (or the total weight of edges) crossing between the clusters. This problem is naturally generalized into the question of generic clustering on a graph: given a graph, can we find a partitioning S 1,..., S k so that k i=1 S i = V and k i=1 S i = so within clusters the graph is dense and between clusters the graph is sparse. A widely successful genre of algorithms for graph partitioning problems is spectral clustering. As evidence, the following technical report giving a tutorial on spectral clustering methods has well over 3,000 citations [1]. Broadly speaking, spectral clustering algorithms use the global information given by the eigenvectors of the adjacency matrix of the graph to embed the graph to a lower dimensional space, i.e., a low rank approximation of the adjacency matrix. Other successful spectral algorithms operate by finding an eigen decomposition of the Laplacian matrix of the graph. One application of spectral clustering of graphs in machine learning is the community detection problem. Here, the assumption is that a graph (for example, the Facebook friend graph) represents relationships between persons, and clusters correspond to communities of friends or related people. A probabilistic model is assumed to generate the graph; the stochastic block model has been the most widely studied for such stochastic process. In the simplest case of two communities, the stochastic block model simply implies that every edge exists with probability p if the two endpoints belong to the same community, or q otherwise, with p > q. The machine learning problem is to recover the community structure given a graph generated by this stochastic process. In an influential result [2], McSherry used spectral clustering techniques to show that community detection can be efficiently accomplished within this simple stochastic block model. Very recently, work has been done to generalize the stochastic block model. Specifically, in [3], the authors work with the generalized stochastic block model; the substantial deviation is that now edges are allowed to have labels (which we will think of as weights), which are assumed to be drawn from an unknown distribution generated by a symmetric kernel on latent variables of the endpoints. The question of whether it is possible to recover the distributions from which the labels are drawn given a partially observed graph is answered in the affirmative, given that the graph is sufficiently 1

2 (logarithmic in V ) dense. However, given the generality of the model, this method is not able to recover the latent variables on the vertices themselves. In this work, we address a model of practical interest in between the basic stochastic block model and the fully generalized model called the labeled stochastic block model: edges will still have labels drawn probabilistically, but we will still assume a standard community structure and attempt to recover this latent state from a graph. This has very recently been attempted in [4]. However, while the authors in [4] focus on semidefinite programming relaxations, we will primarily consider spectral techniques based on reducing the labeled instance to an appropriate instance for the use of McSherry s algorithm. The justification is that McSherry s algorithm and similar simple spectral techniques are easy to implement and robust in practice. We see two immediate applications of understanding community detection in the labeled instance. The first is to infer social network community structure (for analytics, friend reccomendation, etc) when an interaction graph is available instead of just a friend graph, i.e., when the number of interactions in some time period between all pairs of users is available. The second application is discovering network hotspots for diagnosis in a distributed system: given bandwidth information on all connections between servers, what are the communities of servers that are passing excessive information between one another? 2 Methods Formally, we adopt the following labeled stochastic block model for the generation of random labeled graphs G = (V, E) with n nodes. Every vertex v in G belongs to one of k communities; let Ψ : {1,..., n} {1,..., k} denote this community labeling on vertices. p and q denote the intracommunity and inter-community probabilities of drawing an edge respectively; similarly, P and Q represent multinomials on the label set L = {1,..., L} corresponding to probabilities of drawing particular edge labels for intra-community and inter-community edges respectively. A draw of a random graph G from this stochastic process can be generated by the following steps. 1. For every vertex, draw a community from 1 to k according to an arbitrary distribution. 2. For every pair of vertices i, j, the edge (i, j) exists with probability p if i and j are in the same community and q otherwise. 3. For all edges in the graph, we draw a label from L = {1,..., L} according to multinomial distribution P = {p 1,..., p L } if i and j are in the same community and according to multinomial distribution Q = {q 1,..., q L } otherwise. We are given a single draw of a graph Ĝ from the random model and we want to recover the complete labeling Ψ (up to isomorphism) with high probability. We begin by giving McSherry s algorithm for the simpler case where we do not draw edge labels [2]. We assume that the value of k (the number of communities) is known; otherwise it is necessary to search over likely values of k and pick the best clustering. Essentially we will compute the best low rank (specifically, rank k approximation) of the adjacency matrix of Ĝ and cluster vertices according to that low rank approximation. 1: procedure CLUSTER(Ĝ, τ, k) 2: Randomly divide the vertices into sets two sets 3: Let  and ˆB be the columns of Ĝ for the vertices in the first and second set respectively 4: Let P 1 = Project( ˆB, k, τ); let P 2 = Project(Â, k, τ) 5: Let Ĥ be the matrix whose columns are those of (P1)(Â) and (P 2)( ˆB) 6: while U Ĝ unpartitioned do 7: Arbitrarily choose u i U 8: for v Ĝ do 9: 10: if Ĥu i Ĥv 2 τ then ˆΨ(v) = i 11: Return ˆΨ 2

3 In the Cluster algorithm, τ is a small similarity parameter. Ĥ v refers to the column of Ĥ corresponding to vertex v in the transformed adjacency matrix. Project( ˆB, k, τ) returns a projection operator (in the form of a matrix) which is an approximation to the projection onto B. We omit the simple details of this sub procedure here. Theorem 1. For an instance (Ψ, p, q, k) of the unlabeled problem, there is a constant c so that for sufficiently large n if p q log (n/δ) > c p pn then Cluster can recover Ψ with probability 1 δ [2]. We want to generalize from the Cluster algorithm to solve the community detection problem (i.e., to recover Ψ) for labeled graphs. To do this, we introduce a threshold parameter t and reduce a labeled instance to an unlabeled instance by deleting all edges with labels below t and removing labels from all edges with labels at or above t. Some values of t may fail because they produce reduced graphs for which the equivalent values of p and q are not well separated as required by theorem 1. Thus, we introduce a heuristic for selecting a good value of t: we search for a value of t which results in a clustering via McSherry s algorithm for which the expected values of p and q are well separated. We justify the reasoning of this heuristic experimentally. 1: procedure LABELEDCLUSTER(Ĝ, τ, k) 2: for t from 1 to L do 3: Compute Ĝt; delete edges of Ĝ with labels below t and remove all labels 4: Ψ t = Cluster(Ĝt, τ, k) ] 5: p t, q t = E [p t, q t Ĝt, Ψ t 6: t = argmax t (p t q t ) 7: Return Ψ t For example, Figure 1 shows a graph whose labels are represented by colors; each instance from left to right shows the remaining graph passed to the unlabeled Cluster algorithm for increasing values of the threshold parameter t. Green and red vertices correspond to different communities; the intuition of the heuristic is that the easiest instance of this graph for the Cluster algorithm should be the second or third, which should also be the instance where the expected values of p and q are best separated. Figure 1: labeled Cluster example 2.1 Improving the Running Time Naively the algorithm has linear dependence on the value of L. Under certain assumptions on the distributions P and Q, we can improve this to depend only on log L for instances where the label set is large. This arises from a noisy binary search over the label values to use as a threshold, made possible when the likely p q values exhibit concave structure with respect to the threshold. For instance, this occurs when P and Q are well separated symmetric distributions like discrete approximations of Gaussians (Figure 2), or monotone distributions etc. The binary search is noisy because for any given threshold, there is some probability that the clustering algorithm fails and the p and q values computed are nonsense. Consequentially, the binary search should consider median separation of p and q for small neighborhoods of threshold values 3

4 Figure 2: Two discrete approximations of Gaussians (see Figure 3). The size of these neighborhoods considered can be scaled: for fast execution set the neighborhoods to be size 1; for careful execution when the clustering algorithm may fail with non negligible probability, the neighborhoods should be larger. Figure 3: Binary search over groups of points 3 Results and Discussion We implemented the proposed algorithm LabeledCluster and compared the results with a naive implementation of McSherry s algorithm[2] on synthetic graphs. We generated random graphs of different sizes (nodes) with varying intra-cluster and inter-cluster edge probabilities. We generated these random graphs with 2 partitions and the goal is to recover these partitions as closely as possible. We measured the error as the difference between the true partitions and the computed partitions. More specifically, for every pair of nodes, we compare their partitions with the true partitions, i.e., whether they are in the same partitions or not. Based on this, we measure the partition error as the proportions of pairs of nodes that are incorrectly partitioned with respect to the ground truth. Since, we generated the random graphs with known partitions, we consider it to be the ground truth. All of the results are averaged over 10 runs. (a) Without edge labels (b) With edge labels Figure 4: Partition error on synthetic graphs 4

5 We measured the partitioning error of McSherry s algorithm on random graphs for different values of edge probabilities. Here, p denotes the probability of having an edge between two nodes that are in the same partition, and q denotes the probability of having an edge between two nodes that are in the different partitions. We compute error with different p and q as shown on the x axis. As expected, Figure 4a shows that the partitioning error is low when p and q are well separated, and tend to get higher when this separation diminishes. We varied the number of nodes in a graph from 10 to 1000 and found that for a fixed number of partitions, the error reduces with increase in number of nodes. For instance, after fixing p and q to 0.6 and 0.4 respectively, the error on a graph with 10 nodes is 0.47, while, the error on a graph with 1000 nodes is only McSherry s algorithm gives the guarantee of success but does not give any bound on the error when the algorithm fails. On our preliminary results, we found that when there is a good separation between p and q, the algorithm produces partitions with low error even in the cases of failure. Figure 4b shows the comparison of LabeledCluster (dynamic partition) and McSherry s algorithm (fixed partition) on a random graph with 3 edge labels - {l 0, l 1, l 2 } where l 0 represents no edge. The corresponding label probabilities are fixed as p = {0, 0.2, 0.8} and q= {0.5, 0.5, 0)}. As we can see form the figure, the dynamic partitioning scheme performs much better than the fixed partitioning scheme for this specific instance. As explained before, the dynamic partitioning scheme tries all possible thresholds and selects the one that maximizes the distance between intra-cluster and inter-cluster probabilities, which we heuristically take as the best partition. Note, that we show this only for the case of 2 partitions and with a well separated edge probabilities. However, we believe that dynamic partitioning method should give better results compared to the fixed partitioning scheme even in the case of multiple partitions as long as the edge probabilities are well separated. 4 Future Work For future work, it would be interesting to push the implementation of community detection in labeled graphs to real data and or data with large label sets to validate the idea of log L dependent community detection. Furthermore, for real data, we are interested in whether exploiting label information is useful in the sense that it provides better clustering for applications where label information is available compared to other techniques that do not consider label information. Another interesting direction is to use our algorithm to get better clusters than McSherry s algorithms even without prior knowledge of the labels of the graph. Instead of labels we can set weights on the edges taking advantage of the structure of the graph. For example, the weight of an edge (u, v) can be the number of common friends between nodes u and v. So for each edge we find a weight computing the number of common friends between the incident nodes. If we cluster the weights to k levels (the first level contains the highest weights, the last one the lowest weights etc.) then the label of an edge can be defined as the level of its weight. If the number of nodes is sufficiently large, it is possible to have a better separator (in expectation) than the simple McSherry s algortihm. Finally, it is of theoretical interest to ask whether the heuristic of well separated expected p and q values can be justified mathematically as well as intuitively and experimentally. 5 Conclusion The stochastic block model can be fully generalized to describe the generation of nearly arbitrary random matrices with edge labels. However, discovering the latent variables on vertices is not possible in the general model; therefore the classic community detection problem in the generalized model is impossible. However, a slight relaxation of the generalized stochastic block model to a discrete labeled version allows us to model weighted graphs generated according a stochastic process and still solve community detection. We do this by reducing to an unlabeled instance and applying the classic spectral clustering technique from McSherry [2] to recover the latent variables (communities). We have also implemented our algorithm on synthetic data for small label sets and we presented some preliminary results. References [1] Ulrike von Luxburg. A tutorial on spectral clustering. Technical report, Max Planck Institute,

6 [2] F. McSherry. Spectral partitioning of random graphs. In Foundations of Computer Science, FOCS 01, [3] J. Xu, L. Massoulié, and M. Lelarge. Edge Label Inference in Generalized Stochastic Block Models: from Spectral Theory to Impossibility Results. In Conference on Learning Theory, COLT 14, [4] M. Lelarge, L. Massoulié, and J. Xu. Reconstruction in the Labeled Stochastic Block Model. ArXiv e-prints, February

My favorite application using eigenvalues: partitioning and community detection in social networks

My favorite application using eigenvalues: partitioning and community detection in social networks My favorite application using eigenvalues: partitioning and community detection in social networks Will Hobbs February 17, 2013 Abstract Social networks are often organized into families, friendship groups,

More information

Visual Representations for Machine Learning

Visual Representations for Machine Learning Visual Representations for Machine Learning Spectral Clustering and Channel Representations Lecture 1 Spectral Clustering: introduction and confusion Michael Felsberg Klas Nordberg The Spectral Clustering

More information

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 Spectral Clustering Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 What are we going to talk about? Introduction Clustering and

More information

Spectral Clustering X I AO ZE N G + E L HA M TA BA S SI CS E CL A S S P R ESENTATION MA RCH 1 6,

Spectral Clustering X I AO ZE N G + E L HA M TA BA S SI CS E CL A S S P R ESENTATION MA RCH 1 6, Spectral Clustering XIAO ZENG + ELHAM TABASSI CSE 902 CLASS PRESENTATION MARCH 16, 2017 1 Presentation based on 1. Von Luxburg, Ulrike. "A tutorial on spectral clustering." Statistics and computing 17.4

More information

E-Companion: On Styles in Product Design: An Analysis of US. Design Patents

E-Companion: On Styles in Product Design: An Analysis of US. Design Patents E-Companion: On Styles in Product Design: An Analysis of US Design Patents 1 PART A: FORMALIZING THE DEFINITION OF STYLES A.1 Styles as categories of designs of similar form Our task involves categorizing

More information

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components Review Lecture 14 ! PRINCIPAL COMPONENT ANALYSIS Eigenvectors of the covariance matrix are the principal components 1. =cov X Top K principal components are the eigenvectors with K largest eigenvalues

More information

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1

Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanfordedu) February 6, 2018 Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 In the

More information

Exact Recovery in Stochastic Block Models

Exact Recovery in Stochastic Block Models Eact Recovery in Stochastic Block Models Ben Eysenbach May 11, 2016 1 Introduction Graphs are a natural way of capturing interactions between objects, such as phone calls between friends, binding between

More information

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Bryan Poling University of Minnesota Joint work with Gilad Lerman University of Minnesota The Problem of Subspace

More information

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering is one of the fundamental and ubiquitous tasks in exploratory data analysis a first intuition about the

More information

TELCOM2125: Network Science and Analysis

TELCOM2125: Network Science and Analysis School of Information Sciences University of Pittsburgh TELCOM2125: Network Science and Analysis Konstantinos Pelechrinis Spring 2015 2 Part 4: Dividing Networks into Clusters The problem l Graph partitioning

More information

Samuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR

Samuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR Detecting Missing and Spurious Edges in Large, Dense Networks Using Parallel Computing Samuel Coolidge, sam.r.coolidge@gmail.com Dan Simon, des480@nyu.edu Dennis Shasha, shasha@cims.nyu.edu Technical Report

More information

Introduction to spectral clustering

Introduction to spectral clustering Introduction to spectral clustering Vasileios Zografos zografos@isy.liu.se Klas Nordberg klas@isy.liu.se What this course is Basic introduction into the core ideas of spectral clustering Sufficient to

More information

Scalable Clustering of Signed Networks Using Balance Normalized Cut

Scalable Clustering of Signed Networks Using Balance Normalized Cut Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.

More information

Lecture 27: Fast Laplacian Solvers

Lecture 27: Fast Laplacian Solvers Lecture 27: Fast Laplacian Solvers Scribed by Eric Lee, Eston Schweickart, Chengrun Yang November 21, 2017 1 How Fast Laplacian Solvers Work We want to solve Lx = b with L being a Laplacian matrix. Recall

More information

Mining Social Network Graphs

Mining Social Network Graphs Mining Social Network Graphs Analysis of Large Graphs: Community Detection Rafael Ferreira da Silva rafsilva@isi.edu http://rafaelsilva.com Note to other teachers and users of these slides: We would be

More information

Clustering in Networks

Clustering in Networks Clustering in Networks (Spectral Clustering with the Graph Laplacian... a brief introduction) Tom Carter Computer Science CSU Stanislaus http://csustan.csustan.edu/ tom/clustering April 1, 2012 1 Our general

More information

Non-exhaustive, Overlapping k-means

Non-exhaustive, Overlapping k-means Non-exhaustive, Overlapping k-means J. J. Whang, I. S. Dhilon, and D. F. Gleich Teresa Lebair University of Maryland, Baltimore County October 29th, 2015 Teresa Lebair UMBC 1/38 Outline Introduction NEO-K-Means

More information

The clustering in general is the task of grouping a set of objects in such a way that objects

The clustering in general is the task of grouping a set of objects in such a way that objects Spectral Clustering: A Graph Partitioning Point of View Yangzihao Wang Computer Science Department, University of California, Davis yzhwang@ucdavis.edu Abstract This course project provide the basic theory

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Graph based machine learning with applications to media analytics

Graph based machine learning with applications to media analytics Graph based machine learning with applications to media analytics Lei Ding, PhD 9-1-2011 with collaborators at Outline Graph based machine learning Basic structures Algorithms Examples Applications in

More information

RECOVERY OF PARTIALLY OBSERVED DATA APPEARING IN CLUSTERS. Sunrita Poddar, Mathews Jacob

RECOVERY OF PARTIALLY OBSERVED DATA APPEARING IN CLUSTERS. Sunrita Poddar, Mathews Jacob RECOVERY OF PARTIALLY OBSERVED DATA APPEARING IN CLUSTERS Sunrita Poddar, Mathews Jacob Department of Electrical and Computer Engineering The University of Iowa, IA, USA ABSTRACT We propose a matrix completion

More information

Motivation. Technical Background

Motivation. Technical Background Handling Outliers through Agglomerative Clustering with Full Model Maximum Likelihood Estimation, with Application to Flow Cytometry Mark Gordon, Justin Li, Kevin Matzen, Bryce Wiedenbeck Motivation Clustering

More information

Web Structure Mining Community Detection and Evaluation

Web Structure Mining Community Detection and Evaluation Web Structure Mining Community Detection and Evaluation 1 Community Community. It is formed by individuals such that those within a group interact with each other more frequently than with those outside

More information

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM 96 CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM Clustering is the process of combining a set of relevant information in the same group. In this process KM algorithm plays

More information

Explore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan

Explore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan Explore Co-clustering on Job Applications Qingyun Wan SUNet ID:qywan 1 Introduction In the job marketplace, the supply side represents the job postings posted by job posters and the demand side presents

More information

Course Introduction / Review of Fundamentals of Graph Theory

Course Introduction / Review of Fundamentals of Graph Theory Course Introduction / Review of Fundamentals of Graph Theory Hiroki Sayama sayama@binghamton.edu Rise of Network Science (From Barabasi 2010) 2 Network models Many discrete parts involved Classic mean-field

More information

Social Network Analysis

Social Network Analysis Social Network Analysis Mathematics of Networks Manar Mohaisen Department of EEC Engineering Adjacency matrix Network types Edge list Adjacency list Graph representation 2 Adjacency matrix Adjacency matrix

More information

Design of Orthogonal Graph Wavelet Filter Banks

Design of Orthogonal Graph Wavelet Filter Banks Design of Orthogonal Graph Wavelet Filter Banks Xi ZHANG Department of Communication Engineering and Informatics The University of Electro-Communications Chofu-shi, Tokyo, 182-8585 JAPAN E-mail: zhangxi@uec.ac.jp

More information

Topological Classification of Data Sets without an Explicit Metric

Topological Classification of Data Sets without an Explicit Metric Topological Classification of Data Sets without an Explicit Metric Tim Harrington, Andrew Tausz and Guillaume Troianowski December 10, 2008 A contemporary problem in data analysis is understanding the

More information

Cluster Analysis (b) Lijun Zhang

Cluster Analysis (b) Lijun Zhang Cluster Analysis (b) Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Grid-Based and Density-Based Algorithms Graph-Based Algorithms Non-negative Matrix Factorization Cluster Validation Summary

More information

Spectral Methods for Network Community Detection and Graph Partitioning

Spectral Methods for Network Community Detection and Graph Partitioning Spectral Methods for Network Community Detection and Graph Partitioning M. E. J. Newman Department of Physics, University of Michigan Presenters: Yunqi Guo Xueyin Yu Yuanqi Li 1 Outline: Community Detection

More information

Community Detection. Community

Community Detection. Community Community Detection Community In social sciences: Community is formed by individuals such that those within a group interact with each other more frequently than with those outside the group a.k.a. group,

More information

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation

More information

Graph Theory S 1 I 2 I 1 S 2 I 1 I 2

Graph Theory S 1 I 2 I 1 S 2 I 1 I 2 Graph Theory S I I S S I I S Graphs Definition A graph G is a pair consisting of a vertex set V (G), and an edge set E(G) ( ) V (G). x and y are the endpoints of edge e = {x, y}. They are called adjacent

More information

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)

Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html

More information

Spectral Clustering on Handwritten Digits Database

Spectral Clustering on Handwritten Digits Database October 6, 2015 Spectral Clustering on Handwritten Digits Database Danielle dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park Advance

More information

CSCI-B609: A Theorist s Toolkit, Fall 2016 Sept. 6, Firstly let s consider a real world problem: community detection.

CSCI-B609: A Theorist s Toolkit, Fall 2016 Sept. 6, Firstly let s consider a real world problem: community detection. CSCI-B609: A Theorist s Toolkit, Fall 016 Sept. 6, 016 Lecture 03: The Sparsest Cut Problem and Cheeger s Inequality Lecturer: Yuan Zhou Scribe: Xuan Dong We will continue studying the spectral graph theory

More information

Smoothing Dissimilarities for Cluster Analysis: Binary Data and Functional Data

Smoothing Dissimilarities for Cluster Analysis: Binary Data and Functional Data Smoothing Dissimilarities for Cluster Analysis: Binary Data and unctional Data David B. University of South Carolina Department of Statistics Joint work with Zhimin Chen University of South Carolina Current

More information

Graph drawing in spectral layout

Graph drawing in spectral layout Graph drawing in spectral layout Maureen Gallagher Colleen Tygh John Urschel Ludmil Zikatanov Beginning: July 8, 203; Today is: October 2, 203 Introduction Our research focuses on the use of spectral graph

More information

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06 Clustering CS294 Practical Machine Learning Junming Yin 10/09/06 Outline Introduction Unsupervised learning What is clustering? Application Dissimilarity (similarity) of objects Clustering algorithm K-means,

More information

Image Segmentation continued Graph Based Methods

Image Segmentation continued Graph Based Methods Image Segmentation continued Graph Based Methods Previously Images as graphs Fully-connected graph node (vertex) for every pixel link between every pair of pixels, p,q affinity weight w pq for each link

More information

Behavioral Data Mining. Lecture 18 Clustering

Behavioral Data Mining. Lecture 18 Clustering Behavioral Data Mining Lecture 18 Clustering Outline Why? Cluster quality K-means Spectral clustering Generative Models Rationale Given a set {X i } for i = 1,,n, a clustering is a partition of the X i

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

6 Randomized rounding of semidefinite programs

6 Randomized rounding of semidefinite programs 6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can

More information

Asensor network is a collection of interconnected

Asensor network is a collection of interconnected Estimation on Graphs from Relative Measurements PRABIR BAROOAH and JOÃO P. HESPANHA DIGITALVISION DISTRIBUTED ALGORITHMS AND FUNDAMENTAL LIMITS Asensor network is a collection of interconnected nodes that

More information

Statistical Physics of Community Detection

Statistical Physics of Community Detection Statistical Physics of Community Detection Keegan Go (keegango), Kenji Hata (khata) December 8, 2015 1 Introduction Community detection is a key problem in network science. Identifying communities, defined

More information

10701 Machine Learning. Clustering

10701 Machine Learning. Clustering 171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among

More information

Lesson 2 7 Graph Partitioning

Lesson 2 7 Graph Partitioning Lesson 2 7 Graph Partitioning The Graph Partitioning Problem Look at the problem from a different angle: Let s multiply a sparse matrix A by a vector X. Recall the duality between matrices and graphs:

More information

Large-Scale Face Manifold Learning

Large-Scale Face Manifold Learning Large-Scale Face Manifold Learning Sanjiv Kumar Google Research New York, NY * Joint work with A. Talwalkar, H. Rowley and M. Mohri 1 Face Manifold Learning 50 x 50 pixel faces R 2500 50 x 50 pixel random

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Clustering Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1 / 19 Outline

More information

Lecture 3: Art Gallery Problems and Polygon Triangulation

Lecture 3: Art Gallery Problems and Polygon Triangulation EECS 396/496: Computational Geometry Fall 2017 Lecture 3: Art Gallery Problems and Polygon Triangulation Lecturer: Huck Bennett In this lecture, we study the problem of guarding an art gallery (specified

More information

CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning

CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning Parallel sparse matrix-vector product Lay out matrix and vectors by rows y(i) = sum(a(i,j)*x(j)) Only compute terms with A(i,j) 0 P0 P1

More information

The Surprising Power of Belief Propagation

The Surprising Power of Belief Propagation The Surprising Power of Belief Propagation Elchanan Mossel June 12, 2015 Why do you want to know about BP It s a popular algorithm. We will talk abut its analysis. Many open problems. Connections to: Random

More information

What Does Robustness Say About Algorithms?

What Does Robustness Say About Algorithms? What Does Robustness Say About Algorithms? Ankur Moitra (MIT) ICML 2017 Tutorial, August 6 th Let me tell you a story about the tension between sharp thresholds and robustness THE STOCHASTIC BLOCK MODEL

More information

A Probabilistic Approach to the Hough Transform

A Probabilistic Approach to the Hough Transform A Probabilistic Approach to the Hough Transform R S Stephens Computing Devices Eastbourne Ltd, Kings Drive, Eastbourne, East Sussex, BN21 2UE It is shown that there is a strong relationship between the

More information

Intro to Random Graphs and Exponential Random Graph Models

Intro to Random Graphs and Exponential Random Graph Models Intro to Random Graphs and Exponential Random Graph Models Danielle Larcomb University of Denver Danielle Larcomb Random Graphs 1/26 Necessity of Random Graphs The study of complex networks plays an increasingly

More information

Computational Statistics and Mathematics for Cyber Security

Computational Statistics and Mathematics for Cyber Security and Mathematics for Cyber Security David J. Marchette Sept, 0 Acknowledgment: This work funded in part by the NSWC In-House Laboratory Independent Research (ILIR) program. NSWCDD-PN--00 Topics NSWCDD-PN--00

More information

Polygon Partitioning. Lecture03

Polygon Partitioning. Lecture03 1 Polygon Partitioning Lecture03 2 History of Triangulation Algorithms 3 Outline Monotone polygon Triangulation of monotone polygon Trapezoidal decomposition Decomposition in monotone mountain Convex decomposition

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,

More information

MATH 567: Mathematical Techniques in Data

MATH 567: Mathematical Techniques in Data Supervised and unsupervised learning Supervised learning problems: MATH 567: Mathematical Techniques in Data (X, Y ) P (X, Y ). Data Science Clustering I is labelled (input/output) with joint density We

More information

An Introduction to Graph Theory

An Introduction to Graph Theory An Introduction to Graph Theory CIS008-2 Logic and Foundations of Mathematics David Goodwin david.goodwin@perisic.com 12:00, Friday 17 th February 2012 Outline 1 Graphs 2 Paths and cycles 3 Graphs and

More information

Social-Network Graphs

Social-Network Graphs Social-Network Graphs Mining Social Networks Facebook, Google+, Twitter Email Networks, Collaboration Networks Identify communities Similar to clustering Communities usually overlap Identify similarities

More information

What is Computer Vision?

What is Computer Vision? Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras

More information

Spatial-Color Pixel Classification by Spectral Clustering for Color Image Segmentation

Spatial-Color Pixel Classification by Spectral Clustering for Color Image Segmentation 2008 ICTTA Damascus (Syria), April, 2008 Spatial-Color Pixel Classification by Spectral Clustering for Color Image Segmentation Pierre-Alexandre Hébert (LASL) & L. Macaire (LAGIS) Context Summary Segmentation

More information

Distributed minimum spanning tree problem

Distributed minimum spanning tree problem Distributed minimum spanning tree problem Juho-Kustaa Kangas 24th November 2012 Abstract Given a connected weighted undirected graph, the minimum spanning tree problem asks for a spanning subtree with

More information

Some Applications of Graph Bandwidth to Constraint Satisfaction Problems

Some Applications of Graph Bandwidth to Constraint Satisfaction Problems Some Applications of Graph Bandwidth to Constraint Satisfaction Problems Ramin Zabih Computer Science Department Stanford University Stanford, California 94305 Abstract Bandwidth is a fundamental concept

More information

Online Social Networks and Media

Online Social Networks and Media Online Social Networks and Media Absorbing Random Walks Link Prediction Why does the Power Method work? If a matrix R is real and symmetric, it has real eigenvalues and eigenvectors: λ, w, λ 2, w 2,, (λ

More information

Aarti Singh. Machine Learning / Slides Courtesy: Eric Xing, M. Hein & U.V. Luxburg

Aarti Singh. Machine Learning / Slides Courtesy: Eric Xing, M. Hein & U.V. Luxburg Spectral Clustering Aarti Singh Machine Learning 10-701/15-781 Apr 7, 2010 Slides Courtesy: Eric Xing, M. Hein & U.V. Luxburg 1 Data Clustering Graph Clustering Goal: Given data points X1,, Xn and similarities

More information

A Taxonomy of Semi-Supervised Learning Algorithms

A Taxonomy of Semi-Supervised Learning Algorithms A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph

More information

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis

Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Xavier Le Faucheur a, Brani Vidakovic b and Allen Tannenbaum a a School of Electrical and Computer Engineering, b Department of Biomedical

More information

Graph Theory Problem Ideas

Graph Theory Problem Ideas Graph Theory Problem Ideas April 15, 017 Note: Please let me know if you have a problem that you would like me to add to the list! 1 Classification Given a degree sequence d 1,...,d n, let N d1,...,d n

More information

Summary of Raptor Codes

Summary of Raptor Codes Summary of Raptor Codes Tracey Ho October 29, 2003 1 Introduction This summary gives an overview of Raptor Codes, the latest class of codes proposed for reliable multicast in the Digital Fountain model.

More information

Clustering: Classic Methods and Modern Views

Clustering: Classic Methods and Modern Views Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering

More information

Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011

Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 1. Introduction Reddit is one of the most popular online social news websites with millions

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

How Hard Is Inference for Structured Prediction?

How Hard Is Inference for Structured Prediction? How Hard Is Inference for Structured Prediction? Tim Roughgarden (Stanford University) joint work with Amir Globerson (Tel Aviv), David Sontag (NYU), and Cafer Yildirum (NYU) 1 Structured Prediction structured

More information

Analyzing Graph Structure via Linear Measurements.

Analyzing Graph Structure via Linear Measurements. Analyzing Graph Structure via Linear Measurements. Reviewed by Nikhil Desai Introduction Graphs are an integral data structure for many parts of computation. They are highly effective at modeling many

More information

Generalized trace ratio optimization and applications

Generalized trace ratio optimization and applications Generalized trace ratio optimization and applications Mohammed Bellalij, Saïd Hanafi, Rita Macedo and Raca Todosijevic University of Valenciennes, France PGMO Days, 2-4 October 2013 ENSTA ParisTech PGMO

More information

SPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING

SPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING SPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING Alireza Chakeri, Hamidreza Farhidzadeh, Lawrence O. Hall Department of Computer Science and Engineering College of Engineering University of South Florida

More information

A GRAPH FROM THE VIEWPOINT OF ALGEBRAIC TOPOLOGY

A GRAPH FROM THE VIEWPOINT OF ALGEBRAIC TOPOLOGY A GRAPH FROM THE VIEWPOINT OF ALGEBRAIC TOPOLOGY KARL L. STRATOS Abstract. The conventional method of describing a graph as a pair (V, E), where V and E repectively denote the sets of vertices and edges,

More information

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic SEMANTIC COMPUTING Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 23 November 2018 Overview Unsupervised Machine Learning overview Association

More information

Function approximation using RBF network. 10 basis functions and 25 data points.

Function approximation using RBF network. 10 basis functions and 25 data points. 1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Efficient Semi-supervised Spectral Co-clustering with Constraints

Efficient Semi-supervised Spectral Co-clustering with Constraints 2010 IEEE International Conference on Data Mining Efficient Semi-supervised Spectral Co-clustering with Constraints Xiaoxiao Shi, Wei Fan, Philip S. Yu Department of Computer Science, University of Illinois

More information

Clustering Using Graph Connectivity

Clustering Using Graph Connectivity Clustering Using Graph Connectivity Patrick Williams June 3, 010 1 Introduction It is often desirable to group elements of a set into disjoint subsets, based on the similarity between the elements in the

More information

Limitations of Matrix Completion via Trace Norm Minimization

Limitations of Matrix Completion via Trace Norm Minimization Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing

More information

CS281 Section 9: Graph Models and Practical MCMC

CS281 Section 9: Graph Models and Practical MCMC CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs

More information

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro Learning Low-rank Transformations: Algorithms and Applications Qiang Qiu Guillermo Sapiro Motivation Outline Low-rank transform - algorithms and theories Applications Subspace clustering Classification

More information

Problem Definition. Clustering nonlinearly separable data:

Problem Definition. Clustering nonlinearly separable data: Outlines Weighted Graph Cuts without Eigenvectors: A Multilevel Approach (PAMI 2007) User-Guided Large Attributed Graph Clustering with Multiple Sparse Annotations (PAKDD 2016) Problem Definition Clustering

More information

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Descriptive model A descriptive model presents the main features of the data

More information

Comparing the strength of query types in property testing: The case of testing k-colorability

Comparing the strength of query types in property testing: The case of testing k-colorability Comparing the strength of query types in property testing: The case of testing k-colorability Ido Ben-Eliezer Tali Kaufman Michael Krivelevich Dana Ron Abstract We study the power of four query models

More information

Bioinformatics - Lecture 07

Bioinformatics - Lecture 07 Bioinformatics - Lecture 07 Bioinformatics Clusters and networks Martin Saturka http://www.bioplexity.org/lectures/ EBI version 0.4 Creative Commons Attribution-Share Alike 2.5 License Learning on profiles

More information

One-mode Additive Clustering of Multiway Data

One-mode Additive Clustering of Multiway Data One-mode Additive Clustering of Multiway Data Dirk Depril and Iven Van Mechelen KULeuven Tiensestraat 103 3000 Leuven, Belgium (e-mail: dirk.depril@psy.kuleuven.ac.be iven.vanmechelen@psy.kuleuven.ac.be)

More information

Estimating Human Pose in Images. Navraj Singh December 11, 2009

Estimating Human Pose in Images. Navraj Singh December 11, 2009 Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks

More information

On the Approximability of Modularity Clustering

On the Approximability of Modularity Clustering On the Approximability of Modularity Clustering Newman s Community Finding Approach for Social Nets Bhaskar DasGupta Department of Computer Science University of Illinois at Chicago Chicago, IL 60607,

More information

On Dimensionality Reduction of Massive Graphs for Indexing and Retrieval

On Dimensionality Reduction of Massive Graphs for Indexing and Retrieval On Dimensionality Reduction of Massive Graphs for Indexing and Retrieval Charu C. Aggarwal 1, Haixun Wang # IBM T. J. Watson Research Center Hawthorne, NY 153, USA 1 charu@us.ibm.com # Microsoft Research

More information

Chapter 9. Classification and Clustering

Chapter 9. Classification and Clustering Chapter 9 Classification and Clustering Classification and Clustering Classification and clustering are classical pattern recognition and machine learning problems Classification, also referred to as categorization

More information

GRAPH THEORY: AN INTRODUCTION

GRAPH THEORY: AN INTRODUCTION GRAPH THEORY: AN INTRODUCTION BEGINNERS 3/4/2018 1. GRAPHS AND THEIR PROPERTIES A graph G consists of two sets: a set of vertices V, and a set of edges E. A vertex is simply a labeled point. An edge is

More information

Latent Class Modeling as a Probabilistic Extension of K-Means Clustering

Latent Class Modeling as a Probabilistic Extension of K-Means Clustering Latent Class Modeling as a Probabilistic Extension of K-Means Clustering Latent Class Cluster Models According to Kaufman and Rousseeuw (1990), cluster analysis is "the classification of similar objects

More information