Spectral Clustering and Community Detection in Labeled Graphs
|
|
- Ronald Hood
- 5 years ago
- Views:
Transcription
1 Spectral Clustering and Community Detection in Labeled Graphs Brandon Fain, Stavros Sintos, Nisarg Raval Machine Learning (CompSci 571D / STA 561D) December 7, 2015 {btfain, nisarg, ssintos} at cs.duke.edu Abstract We study spectral clustering techniques to learn community structures in labeled random graphs where edge labels from a label set L = {1,..., L} are drawn according to discrete probability distributions parametrized by community membership of the two end-nodes of the edge. This is a strict generalization of the standard stochastic block model for community detection. 1 Introduction and Related Work Graph partitioning is a fundamental problem in computer science first posed as the minimum cut problem: given a graph G = (V, E), one tries to find a partitioning of the vertices into two clusters S 1, S 2 such that S 1 S 2 = V and S 1 S 2 = so as to minimize the number of edges (or the total weight of edges) crossing between the clusters. This problem is naturally generalized into the question of generic clustering on a graph: given a graph, can we find a partitioning S 1,..., S k so that k i=1 S i = V and k i=1 S i = so within clusters the graph is dense and between clusters the graph is sparse. A widely successful genre of algorithms for graph partitioning problems is spectral clustering. As evidence, the following technical report giving a tutorial on spectral clustering methods has well over 3,000 citations [1]. Broadly speaking, spectral clustering algorithms use the global information given by the eigenvectors of the adjacency matrix of the graph to embed the graph to a lower dimensional space, i.e., a low rank approximation of the adjacency matrix. Other successful spectral algorithms operate by finding an eigen decomposition of the Laplacian matrix of the graph. One application of spectral clustering of graphs in machine learning is the community detection problem. Here, the assumption is that a graph (for example, the Facebook friend graph) represents relationships between persons, and clusters correspond to communities of friends or related people. A probabilistic model is assumed to generate the graph; the stochastic block model has been the most widely studied for such stochastic process. In the simplest case of two communities, the stochastic block model simply implies that every edge exists with probability p if the two endpoints belong to the same community, or q otherwise, with p > q. The machine learning problem is to recover the community structure given a graph generated by this stochastic process. In an influential result [2], McSherry used spectral clustering techniques to show that community detection can be efficiently accomplished within this simple stochastic block model. Very recently, work has been done to generalize the stochastic block model. Specifically, in [3], the authors work with the generalized stochastic block model; the substantial deviation is that now edges are allowed to have labels (which we will think of as weights), which are assumed to be drawn from an unknown distribution generated by a symmetric kernel on latent variables of the endpoints. The question of whether it is possible to recover the distributions from which the labels are drawn given a partially observed graph is answered in the affirmative, given that the graph is sufficiently 1
2 (logarithmic in V ) dense. However, given the generality of the model, this method is not able to recover the latent variables on the vertices themselves. In this work, we address a model of practical interest in between the basic stochastic block model and the fully generalized model called the labeled stochastic block model: edges will still have labels drawn probabilistically, but we will still assume a standard community structure and attempt to recover this latent state from a graph. This has very recently been attempted in [4]. However, while the authors in [4] focus on semidefinite programming relaxations, we will primarily consider spectral techniques based on reducing the labeled instance to an appropriate instance for the use of McSherry s algorithm. The justification is that McSherry s algorithm and similar simple spectral techniques are easy to implement and robust in practice. We see two immediate applications of understanding community detection in the labeled instance. The first is to infer social network community structure (for analytics, friend reccomendation, etc) when an interaction graph is available instead of just a friend graph, i.e., when the number of interactions in some time period between all pairs of users is available. The second application is discovering network hotspots for diagnosis in a distributed system: given bandwidth information on all connections between servers, what are the communities of servers that are passing excessive information between one another? 2 Methods Formally, we adopt the following labeled stochastic block model for the generation of random labeled graphs G = (V, E) with n nodes. Every vertex v in G belongs to one of k communities; let Ψ : {1,..., n} {1,..., k} denote this community labeling on vertices. p and q denote the intracommunity and inter-community probabilities of drawing an edge respectively; similarly, P and Q represent multinomials on the label set L = {1,..., L} corresponding to probabilities of drawing particular edge labels for intra-community and inter-community edges respectively. A draw of a random graph G from this stochastic process can be generated by the following steps. 1. For every vertex, draw a community from 1 to k according to an arbitrary distribution. 2. For every pair of vertices i, j, the edge (i, j) exists with probability p if i and j are in the same community and q otherwise. 3. For all edges in the graph, we draw a label from L = {1,..., L} according to multinomial distribution P = {p 1,..., p L } if i and j are in the same community and according to multinomial distribution Q = {q 1,..., q L } otherwise. We are given a single draw of a graph Ĝ from the random model and we want to recover the complete labeling Ψ (up to isomorphism) with high probability. We begin by giving McSherry s algorithm for the simpler case where we do not draw edge labels [2]. We assume that the value of k (the number of communities) is known; otherwise it is necessary to search over likely values of k and pick the best clustering. Essentially we will compute the best low rank (specifically, rank k approximation) of the adjacency matrix of Ĝ and cluster vertices according to that low rank approximation. 1: procedure CLUSTER(Ĝ, τ, k) 2: Randomly divide the vertices into sets two sets 3: Let  and ˆB be the columns of Ĝ for the vertices in the first and second set respectively 4: Let P 1 = Project( ˆB, k, τ); let P 2 = Project(Â, k, τ) 5: Let Ĥ be the matrix whose columns are those of (P1)(Â) and (P 2)( ˆB) 6: while U Ĝ unpartitioned do 7: Arbitrarily choose u i U 8: for v Ĝ do 9: 10: if Ĥu i Ĥv 2 τ then ˆΨ(v) = i 11: Return ˆΨ 2
3 In the Cluster algorithm, τ is a small similarity parameter. Ĥ v refers to the column of Ĥ corresponding to vertex v in the transformed adjacency matrix. Project( ˆB, k, τ) returns a projection operator (in the form of a matrix) which is an approximation to the projection onto B. We omit the simple details of this sub procedure here. Theorem 1. For an instance (Ψ, p, q, k) of the unlabeled problem, there is a constant c so that for sufficiently large n if p q log (n/δ) > c p pn then Cluster can recover Ψ with probability 1 δ [2]. We want to generalize from the Cluster algorithm to solve the community detection problem (i.e., to recover Ψ) for labeled graphs. To do this, we introduce a threshold parameter t and reduce a labeled instance to an unlabeled instance by deleting all edges with labels below t and removing labels from all edges with labels at or above t. Some values of t may fail because they produce reduced graphs for which the equivalent values of p and q are not well separated as required by theorem 1. Thus, we introduce a heuristic for selecting a good value of t: we search for a value of t which results in a clustering via McSherry s algorithm for which the expected values of p and q are well separated. We justify the reasoning of this heuristic experimentally. 1: procedure LABELEDCLUSTER(Ĝ, τ, k) 2: for t from 1 to L do 3: Compute Ĝt; delete edges of Ĝ with labels below t and remove all labels 4: Ψ t = Cluster(Ĝt, τ, k) ] 5: p t, q t = E [p t, q t Ĝt, Ψ t 6: t = argmax t (p t q t ) 7: Return Ψ t For example, Figure 1 shows a graph whose labels are represented by colors; each instance from left to right shows the remaining graph passed to the unlabeled Cluster algorithm for increasing values of the threshold parameter t. Green and red vertices correspond to different communities; the intuition of the heuristic is that the easiest instance of this graph for the Cluster algorithm should be the second or third, which should also be the instance where the expected values of p and q are best separated. Figure 1: labeled Cluster example 2.1 Improving the Running Time Naively the algorithm has linear dependence on the value of L. Under certain assumptions on the distributions P and Q, we can improve this to depend only on log L for instances where the label set is large. This arises from a noisy binary search over the label values to use as a threshold, made possible when the likely p q values exhibit concave structure with respect to the threshold. For instance, this occurs when P and Q are well separated symmetric distributions like discrete approximations of Gaussians (Figure 2), or monotone distributions etc. The binary search is noisy because for any given threshold, there is some probability that the clustering algorithm fails and the p and q values computed are nonsense. Consequentially, the binary search should consider median separation of p and q for small neighborhoods of threshold values 3
4 Figure 2: Two discrete approximations of Gaussians (see Figure 3). The size of these neighborhoods considered can be scaled: for fast execution set the neighborhoods to be size 1; for careful execution when the clustering algorithm may fail with non negligible probability, the neighborhoods should be larger. Figure 3: Binary search over groups of points 3 Results and Discussion We implemented the proposed algorithm LabeledCluster and compared the results with a naive implementation of McSherry s algorithm[2] on synthetic graphs. We generated random graphs of different sizes (nodes) with varying intra-cluster and inter-cluster edge probabilities. We generated these random graphs with 2 partitions and the goal is to recover these partitions as closely as possible. We measured the error as the difference between the true partitions and the computed partitions. More specifically, for every pair of nodes, we compare their partitions with the true partitions, i.e., whether they are in the same partitions or not. Based on this, we measure the partition error as the proportions of pairs of nodes that are incorrectly partitioned with respect to the ground truth. Since, we generated the random graphs with known partitions, we consider it to be the ground truth. All of the results are averaged over 10 runs. (a) Without edge labels (b) With edge labels Figure 4: Partition error on synthetic graphs 4
5 We measured the partitioning error of McSherry s algorithm on random graphs for different values of edge probabilities. Here, p denotes the probability of having an edge between two nodes that are in the same partition, and q denotes the probability of having an edge between two nodes that are in the different partitions. We compute error with different p and q as shown on the x axis. As expected, Figure 4a shows that the partitioning error is low when p and q are well separated, and tend to get higher when this separation diminishes. We varied the number of nodes in a graph from 10 to 1000 and found that for a fixed number of partitions, the error reduces with increase in number of nodes. For instance, after fixing p and q to 0.6 and 0.4 respectively, the error on a graph with 10 nodes is 0.47, while, the error on a graph with 1000 nodes is only McSherry s algorithm gives the guarantee of success but does not give any bound on the error when the algorithm fails. On our preliminary results, we found that when there is a good separation between p and q, the algorithm produces partitions with low error even in the cases of failure. Figure 4b shows the comparison of LabeledCluster (dynamic partition) and McSherry s algorithm (fixed partition) on a random graph with 3 edge labels - {l 0, l 1, l 2 } where l 0 represents no edge. The corresponding label probabilities are fixed as p = {0, 0.2, 0.8} and q= {0.5, 0.5, 0)}. As we can see form the figure, the dynamic partitioning scheme performs much better than the fixed partitioning scheme for this specific instance. As explained before, the dynamic partitioning scheme tries all possible thresholds and selects the one that maximizes the distance between intra-cluster and inter-cluster probabilities, which we heuristically take as the best partition. Note, that we show this only for the case of 2 partitions and with a well separated edge probabilities. However, we believe that dynamic partitioning method should give better results compared to the fixed partitioning scheme even in the case of multiple partitions as long as the edge probabilities are well separated. 4 Future Work For future work, it would be interesting to push the implementation of community detection in labeled graphs to real data and or data with large label sets to validate the idea of log L dependent community detection. Furthermore, for real data, we are interested in whether exploiting label information is useful in the sense that it provides better clustering for applications where label information is available compared to other techniques that do not consider label information. Another interesting direction is to use our algorithm to get better clusters than McSherry s algorithms even without prior knowledge of the labels of the graph. Instead of labels we can set weights on the edges taking advantage of the structure of the graph. For example, the weight of an edge (u, v) can be the number of common friends between nodes u and v. So for each edge we find a weight computing the number of common friends between the incident nodes. If we cluster the weights to k levels (the first level contains the highest weights, the last one the lowest weights etc.) then the label of an edge can be defined as the level of its weight. If the number of nodes is sufficiently large, it is possible to have a better separator (in expectation) than the simple McSherry s algortihm. Finally, it is of theoretical interest to ask whether the heuristic of well separated expected p and q values can be justified mathematically as well as intuitively and experimentally. 5 Conclusion The stochastic block model can be fully generalized to describe the generation of nearly arbitrary random matrices with edge labels. However, discovering the latent variables on vertices is not possible in the general model; therefore the classic community detection problem in the generalized model is impossible. However, a slight relaxation of the generalized stochastic block model to a discrete labeled version allows us to model weighted graphs generated according a stochastic process and still solve community detection. We do this by reducing to an unlabeled instance and applying the classic spectral clustering technique from McSherry [2] to recover the latent variables (communities). We have also implemented our algorithm on synthetic data for small label sets and we presented some preliminary results. References [1] Ulrike von Luxburg. A tutorial on spectral clustering. Technical report, Max Planck Institute,
6 [2] F. McSherry. Spectral partitioning of random graphs. In Foundations of Computer Science, FOCS 01, [3] J. Xu, L. Massoulié, and M. Lelarge. Edge Label Inference in Generalized Stochastic Block Models: from Spectral Theory to Impossibility Results. In Conference on Learning Theory, COLT 14, [4] M. Lelarge, L. Massoulié, and J. Xu. Reconstruction in the Labeled Stochastic Block Model. ArXiv e-prints, February
My favorite application using eigenvalues: partitioning and community detection in social networks
My favorite application using eigenvalues: partitioning and community detection in social networks Will Hobbs February 17, 2013 Abstract Social networks are often organized into families, friendship groups,
More informationVisual Representations for Machine Learning
Visual Representations for Machine Learning Spectral Clustering and Channel Representations Lecture 1 Spectral Clustering: introduction and confusion Michael Felsberg Klas Nordberg The Spectral Clustering
More informationSpectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014
Spectral Clustering Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 What are we going to talk about? Introduction Clustering and
More informationSpectral Clustering X I AO ZE N G + E L HA M TA BA S SI CS E CL A S S P R ESENTATION MA RCH 1 6,
Spectral Clustering XIAO ZENG + ELHAM TABASSI CSE 902 CLASS PRESENTATION MARCH 16, 2017 1 Presentation based on 1. Von Luxburg, Ulrike. "A tutorial on spectral clustering." Statistics and computing 17.4
More informationE-Companion: On Styles in Product Design: An Analysis of US. Design Patents
E-Companion: On Styles in Product Design: An Analysis of US Design Patents 1 PART A: FORMALIZING THE DEFINITION OF STYLES A.1 Styles as categories of designs of similar form Our task involves categorizing
More information( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components
Review Lecture 14 ! PRINCIPAL COMPONENT ANALYSIS Eigenvectors of the covariance matrix are the principal components 1. =cov X Top K principal components are the eigenvectors with K largest eigenvalues
More informationLecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1
CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanfordedu) February 6, 2018 Lecture 9 - Matrix Multiplication Equivalences and Spectral Graph Theory 1 In the
More informationExact Recovery in Stochastic Block Models
Eact Recovery in Stochastic Block Models Ben Eysenbach May 11, 2016 1 Introduction Graphs are a natural way of capturing interactions between objects, such as phone calls between friends, binding between
More informationSubspace Clustering with Global Dimension Minimization And Application to Motion Segmentation
Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation Bryan Poling University of Minnesota Joint work with Gilad Lerman University of Minnesota The Problem of Subspace
More informationClustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic
Clustering SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering is one of the fundamental and ubiquitous tasks in exploratory data analysis a first intuition about the
More informationTELCOM2125: Network Science and Analysis
School of Information Sciences University of Pittsburgh TELCOM2125: Network Science and Analysis Konstantinos Pelechrinis Spring 2015 2 Part 4: Dividing Networks into Clusters The problem l Graph partitioning
More informationSamuel Coolidge, Dan Simon, Dennis Shasha, Technical Report NYU/CIMS/TR
Detecting Missing and Spurious Edges in Large, Dense Networks Using Parallel Computing Samuel Coolidge, sam.r.coolidge@gmail.com Dan Simon, des480@nyu.edu Dennis Shasha, shasha@cims.nyu.edu Technical Report
More informationIntroduction to spectral clustering
Introduction to spectral clustering Vasileios Zografos zografos@isy.liu.se Klas Nordberg klas@isy.liu.se What this course is Basic introduction into the core ideas of spectral clustering Sufficient to
More informationScalable Clustering of Signed Networks Using Balance Normalized Cut
Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.
More informationLecture 27: Fast Laplacian Solvers
Lecture 27: Fast Laplacian Solvers Scribed by Eric Lee, Eston Schweickart, Chengrun Yang November 21, 2017 1 How Fast Laplacian Solvers Work We want to solve Lx = b with L being a Laplacian matrix. Recall
More informationMining Social Network Graphs
Mining Social Network Graphs Analysis of Large Graphs: Community Detection Rafael Ferreira da Silva rafsilva@isi.edu http://rafaelsilva.com Note to other teachers and users of these slides: We would be
More informationClustering in Networks
Clustering in Networks (Spectral Clustering with the Graph Laplacian... a brief introduction) Tom Carter Computer Science CSU Stanislaus http://csustan.csustan.edu/ tom/clustering April 1, 2012 1 Our general
More informationNon-exhaustive, Overlapping k-means
Non-exhaustive, Overlapping k-means J. J. Whang, I. S. Dhilon, and D. F. Gleich Teresa Lebair University of Maryland, Baltimore County October 29th, 2015 Teresa Lebair UMBC 1/38 Outline Introduction NEO-K-Means
More informationThe clustering in general is the task of grouping a set of objects in such a way that objects
Spectral Clustering: A Graph Partitioning Point of View Yangzihao Wang Computer Science Department, University of California, Davis yzhwang@ucdavis.edu Abstract This course project provide the basic theory
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-
More informationGraph based machine learning with applications to media analytics
Graph based machine learning with applications to media analytics Lei Ding, PhD 9-1-2011 with collaborators at Outline Graph based machine learning Basic structures Algorithms Examples Applications in
More informationRECOVERY OF PARTIALLY OBSERVED DATA APPEARING IN CLUSTERS. Sunrita Poddar, Mathews Jacob
RECOVERY OF PARTIALLY OBSERVED DATA APPEARING IN CLUSTERS Sunrita Poddar, Mathews Jacob Department of Electrical and Computer Engineering The University of Iowa, IA, USA ABSTRACT We propose a matrix completion
More informationMotivation. Technical Background
Handling Outliers through Agglomerative Clustering with Full Model Maximum Likelihood Estimation, with Application to Flow Cytometry Mark Gordon, Justin Li, Kevin Matzen, Bryce Wiedenbeck Motivation Clustering
More informationWeb Structure Mining Community Detection and Evaluation
Web Structure Mining Community Detection and Evaluation 1 Community Community. It is formed by individuals such that those within a group interact with each other more frequently than with those outside
More informationCHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM
96 CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM Clustering is the process of combining a set of relevant information in the same group. In this process KM algorithm plays
More informationExplore Co-clustering on Job Applications. Qingyun Wan SUNet ID:qywan
Explore Co-clustering on Job Applications Qingyun Wan SUNet ID:qywan 1 Introduction In the job marketplace, the supply side represents the job postings posted by job posters and the demand side presents
More informationCourse Introduction / Review of Fundamentals of Graph Theory
Course Introduction / Review of Fundamentals of Graph Theory Hiroki Sayama sayama@binghamton.edu Rise of Network Science (From Barabasi 2010) 2 Network models Many discrete parts involved Classic mean-field
More informationSocial Network Analysis
Social Network Analysis Mathematics of Networks Manar Mohaisen Department of EEC Engineering Adjacency matrix Network types Edge list Adjacency list Graph representation 2 Adjacency matrix Adjacency matrix
More informationDesign of Orthogonal Graph Wavelet Filter Banks
Design of Orthogonal Graph Wavelet Filter Banks Xi ZHANG Department of Communication Engineering and Informatics The University of Electro-Communications Chofu-shi, Tokyo, 182-8585 JAPAN E-mail: zhangxi@uec.ac.jp
More informationTopological Classification of Data Sets without an Explicit Metric
Topological Classification of Data Sets without an Explicit Metric Tim Harrington, Andrew Tausz and Guillaume Troianowski December 10, 2008 A contemporary problem in data analysis is understanding the
More informationCluster Analysis (b) Lijun Zhang
Cluster Analysis (b) Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Grid-Based and Density-Based Algorithms Graph-Based Algorithms Non-negative Matrix Factorization Cluster Validation Summary
More informationSpectral Methods for Network Community Detection and Graph Partitioning
Spectral Methods for Network Community Detection and Graph Partitioning M. E. J. Newman Department of Physics, University of Michigan Presenters: Yunqi Guo Xueyin Yu Yuanqi Li 1 Outline: Community Detection
More informationCommunity Detection. Community
Community Detection Community In social sciences: Community is formed by individuals such that those within a group interact with each other more frequently than with those outside the group a.k.a. group,
More informationClustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation
More informationGraph Theory S 1 I 2 I 1 S 2 I 1 I 2
Graph Theory S I I S S I I S Graphs Definition A graph G is a pair consisting of a vertex set V (G), and an edge set E(G) ( ) V (G). x and y are the endpoints of edge e = {x, y}. They are called adjacent
More informationBiometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong)
Biometrics Technology: Image Processing & Pattern Recognition (by Dr. Dickson Tong) References: [1] http://homepages.inf.ed.ac.uk/rbf/hipr2/index.htm [2] http://www.cs.wisc.edu/~dyer/cs540/notes/vision.html
More informationSpectral Clustering on Handwritten Digits Database
October 6, 2015 Spectral Clustering on Handwritten Digits Database Danielle dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park Advance
More informationCSCI-B609: A Theorist s Toolkit, Fall 2016 Sept. 6, Firstly let s consider a real world problem: community detection.
CSCI-B609: A Theorist s Toolkit, Fall 016 Sept. 6, 016 Lecture 03: The Sparsest Cut Problem and Cheeger s Inequality Lecturer: Yuan Zhou Scribe: Xuan Dong We will continue studying the spectral graph theory
More informationSmoothing Dissimilarities for Cluster Analysis: Binary Data and Functional Data
Smoothing Dissimilarities for Cluster Analysis: Binary Data and unctional Data David B. University of South Carolina Department of Statistics Joint work with Zhimin Chen University of South Carolina Current
More informationGraph drawing in spectral layout
Graph drawing in spectral layout Maureen Gallagher Colleen Tygh John Urschel Ludmil Zikatanov Beginning: July 8, 203; Today is: October 2, 203 Introduction Our research focuses on the use of spectral graph
More informationClustering. CS294 Practical Machine Learning Junming Yin 10/09/06
Clustering CS294 Practical Machine Learning Junming Yin 10/09/06 Outline Introduction Unsupervised learning What is clustering? Application Dissimilarity (similarity) of objects Clustering algorithm K-means,
More informationImage Segmentation continued Graph Based Methods
Image Segmentation continued Graph Based Methods Previously Images as graphs Fully-connected graph node (vertex) for every pixel link between every pair of pixels, p,q affinity weight w pq for each link
More informationBehavioral Data Mining. Lecture 18 Clustering
Behavioral Data Mining Lecture 18 Clustering Outline Why? Cluster quality K-means Spectral clustering Generative Models Rationale Given a set {X i } for i = 1,,n, a clustering is a partition of the X i
More informationRandom projection for non-gaussian mixture models
Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,
More information6 Randomized rounding of semidefinite programs
6 Randomized rounding of semidefinite programs We now turn to a new tool which gives substantially improved performance guarantees for some problems We now show how nonlinear programming relaxations can
More informationAsensor network is a collection of interconnected
Estimation on Graphs from Relative Measurements PRABIR BAROOAH and JOÃO P. HESPANHA DIGITALVISION DISTRIBUTED ALGORITHMS AND FUNDAMENTAL LIMITS Asensor network is a collection of interconnected nodes that
More informationStatistical Physics of Community Detection
Statistical Physics of Community Detection Keegan Go (keegango), Kenji Hata (khata) December 8, 2015 1 Introduction Community detection is a key problem in network science. Identifying communities, defined
More information10701 Machine Learning. Clustering
171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among
More informationLesson 2 7 Graph Partitioning
Lesson 2 7 Graph Partitioning The Graph Partitioning Problem Look at the problem from a different angle: Let s multiply a sparse matrix A by a vector X. Recall the duality between matrices and graphs:
More informationLarge-Scale Face Manifold Learning
Large-Scale Face Manifold Learning Sanjiv Kumar Google Research New York, NY * Joint work with A. Talwalkar, H. Rowley and M. Mohri 1 Face Manifold Learning 50 x 50 pixel faces R 2500 50 x 50 pixel random
More informationIntroduction to Machine Learning
Introduction to Machine Learning Clustering Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1 / 19 Outline
More informationLecture 3: Art Gallery Problems and Polygon Triangulation
EECS 396/496: Computational Geometry Fall 2017 Lecture 3: Art Gallery Problems and Polygon Triangulation Lecturer: Huck Bennett In this lecture, we study the problem of guarding an art gallery (specified
More informationCS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning
CS 140: Sparse Matrix-Vector Multiplication and Graph Partitioning Parallel sparse matrix-vector product Lay out matrix and vectors by rows y(i) = sum(a(i,j)*x(j)) Only compute terms with A(i,j) 0 P0 P1
More informationThe Surprising Power of Belief Propagation
The Surprising Power of Belief Propagation Elchanan Mossel June 12, 2015 Why do you want to know about BP It s a popular algorithm. We will talk abut its analysis. Many open problems. Connections to: Random
More informationWhat Does Robustness Say About Algorithms?
What Does Robustness Say About Algorithms? Ankur Moitra (MIT) ICML 2017 Tutorial, August 6 th Let me tell you a story about the tension between sharp thresholds and robustness THE STOCHASTIC BLOCK MODEL
More informationA Probabilistic Approach to the Hough Transform
A Probabilistic Approach to the Hough Transform R S Stephens Computing Devices Eastbourne Ltd, Kings Drive, Eastbourne, East Sussex, BN21 2UE It is shown that there is a strong relationship between the
More informationIntro to Random Graphs and Exponential Random Graph Models
Intro to Random Graphs and Exponential Random Graph Models Danielle Larcomb University of Denver Danielle Larcomb Random Graphs 1/26 Necessity of Random Graphs The study of complex networks plays an increasingly
More informationComputational Statistics and Mathematics for Cyber Security
and Mathematics for Cyber Security David J. Marchette Sept, 0 Acknowledgment: This work funded in part by the NSWC In-House Laboratory Independent Research (ILIR) program. NSWCDD-PN--00 Topics NSWCDD-PN--00
More informationPolygon Partitioning. Lecture03
1 Polygon Partitioning Lecture03 2 History of Triangulation Algorithms 3 Outline Monotone polygon Triangulation of monotone polygon Trapezoidal decomposition Decomposition in monotone mountain Convex decomposition
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,
More informationMATH 567: Mathematical Techniques in Data
Supervised and unsupervised learning Supervised learning problems: MATH 567: Mathematical Techniques in Data (X, Y ) P (X, Y ). Data Science Clustering I is labelled (input/output) with joint density We
More informationAn Introduction to Graph Theory
An Introduction to Graph Theory CIS008-2 Logic and Foundations of Mathematics David Goodwin david.goodwin@perisic.com 12:00, Friday 17 th February 2012 Outline 1 Graphs 2 Paths and cycles 3 Graphs and
More informationSocial-Network Graphs
Social-Network Graphs Mining Social Networks Facebook, Google+, Twitter Email Networks, Collaboration Networks Identify communities Similar to clustering Communities usually overlap Identify similarities
More informationWhat is Computer Vision?
Perceptual Grouping in Computer Vision Gérard Medioni University of Southern California What is Computer Vision? Computer Vision Attempt to emulate Human Visual System Perceive visual stimuli with cameras
More informationSpatial-Color Pixel Classification by Spectral Clustering for Color Image Segmentation
2008 ICTTA Damascus (Syria), April, 2008 Spatial-Color Pixel Classification by Spectral Clustering for Color Image Segmentation Pierre-Alexandre Hébert (LASL) & L. Macaire (LAGIS) Context Summary Segmentation
More informationDistributed minimum spanning tree problem
Distributed minimum spanning tree problem Juho-Kustaa Kangas 24th November 2012 Abstract Given a connected weighted undirected graph, the minimum spanning tree problem asks for a spanning subtree with
More informationSome Applications of Graph Bandwidth to Constraint Satisfaction Problems
Some Applications of Graph Bandwidth to Constraint Satisfaction Problems Ramin Zabih Computer Science Department Stanford University Stanford, California 94305 Abstract Bandwidth is a fundamental concept
More informationOnline Social Networks and Media
Online Social Networks and Media Absorbing Random Walks Link Prediction Why does the Power Method work? If a matrix R is real and symmetric, it has real eigenvalues and eigenvectors: λ, w, λ 2, w 2,, (λ
More informationAarti Singh. Machine Learning / Slides Courtesy: Eric Xing, M. Hein & U.V. Luxburg
Spectral Clustering Aarti Singh Machine Learning 10-701/15-781 Apr 7, 2010 Slides Courtesy: Eric Xing, M. Hein & U.V. Luxburg 1 Data Clustering Graph Clustering Goal: Given data points X1,, Xn and similarities
More informationA Taxonomy of Semi-Supervised Learning Algorithms
A Taxonomy of Semi-Supervised Learning Algorithms Olivier Chapelle Max Planck Institute for Biological Cybernetics December 2005 Outline 1 Introduction 2 Generative models 3 Low density separation 4 Graph
More informationBayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis
Bayesian Spherical Wavelet Shrinkage: Applications to Shape Analysis Xavier Le Faucheur a, Brani Vidakovic b and Allen Tannenbaum a a School of Electrical and Computer Engineering, b Department of Biomedical
More informationGraph Theory Problem Ideas
Graph Theory Problem Ideas April 15, 017 Note: Please let me know if you have a problem that you would like me to add to the list! 1 Classification Given a degree sequence d 1,...,d n, let N d1,...,d n
More informationSummary of Raptor Codes
Summary of Raptor Codes Tracey Ho October 29, 2003 1 Introduction This summary gives an overview of Raptor Codes, the latest class of codes proposed for reliable multicast in the Digital Fountain model.
More informationClustering: Classic Methods and Modern Views
Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering
More informationReddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011
Reddit Recommendation System Daniel Poon, Yu Wu, David (Qifan) Zhang CS229, Stanford University December 11 th, 2011 1. Introduction Reddit is one of the most popular online social news websites with millions
More informationNote Set 4: Finite Mixture Models and the EM Algorithm
Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for
More informationHow Hard Is Inference for Structured Prediction?
How Hard Is Inference for Structured Prediction? Tim Roughgarden (Stanford University) joint work with Amir Globerson (Tel Aviv), David Sontag (NYU), and Cafer Yildirum (NYU) 1 Structured Prediction structured
More informationAnalyzing Graph Structure via Linear Measurements.
Analyzing Graph Structure via Linear Measurements. Reviewed by Nikhil Desai Introduction Graphs are an integral data structure for many parts of computation. They are highly effective at modeling many
More informationGeneralized trace ratio optimization and applications
Generalized trace ratio optimization and applications Mohammed Bellalij, Saïd Hanafi, Rita Macedo and Raca Todosijevic University of Valenciennes, France PGMO Days, 2-4 October 2013 ENSTA ParisTech PGMO
More informationSPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING
SPECTRAL SPARSIFICATION IN SPECTRAL CLUSTERING Alireza Chakeri, Hamidreza Farhidzadeh, Lawrence O. Hall Department of Computer Science and Engineering College of Engineering University of South Florida
More informationA GRAPH FROM THE VIEWPOINT OF ALGEBRAIC TOPOLOGY
A GRAPH FROM THE VIEWPOINT OF ALGEBRAIC TOPOLOGY KARL L. STRATOS Abstract. The conventional method of describing a graph as a pair (V, E), where V and E repectively denote the sets of vertices and edges,
More informationLecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic
SEMANTIC COMPUTING Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 23 November 2018 Overview Unsupervised Machine Learning overview Association
More informationFunction approximation using RBF network. 10 basis functions and 25 data points.
1 Function approximation using RBF network F (x j ) = m 1 w i ϕ( x j t i ) i=1 j = 1... N, m 1 = 10, N = 25 10 basis functions and 25 data points. Basis function centers are plotted with circles and data
More information10-701/15-781, Fall 2006, Final
-7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly
More informationEfficient Semi-supervised Spectral Co-clustering with Constraints
2010 IEEE International Conference on Data Mining Efficient Semi-supervised Spectral Co-clustering with Constraints Xiaoxiao Shi, Wei Fan, Philip S. Yu Department of Computer Science, University of Illinois
More informationClustering Using Graph Connectivity
Clustering Using Graph Connectivity Patrick Williams June 3, 010 1 Introduction It is often desirable to group elements of a set into disjoint subsets, based on the similarity between the elements in the
More informationLimitations of Matrix Completion via Trace Norm Minimization
Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing
More informationCS281 Section 9: Graph Models and Practical MCMC
CS281 Section 9: Graph Models and Practical MCMC Scott Linderman November 11, 213 Now that we have a few MCMC inference algorithms in our toolbox, let s try them out on some random graph models. Graphs
More informationLearning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro
Learning Low-rank Transformations: Algorithms and Applications Qiang Qiu Guillermo Sapiro Motivation Outline Low-rank transform - algorithms and theories Applications Subspace clustering Classification
More informationProblem Definition. Clustering nonlinearly separable data:
Outlines Weighted Graph Cuts without Eigenvectors: A Multilevel Approach (PAMI 2007) User-Guided Large Attributed Graph Clustering with Multiple Sparse Annotations (PAKDD 2016) Problem Definition Clustering
More informationData Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University
Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Descriptive model A descriptive model presents the main features of the data
More informationComparing the strength of query types in property testing: The case of testing k-colorability
Comparing the strength of query types in property testing: The case of testing k-colorability Ido Ben-Eliezer Tali Kaufman Michael Krivelevich Dana Ron Abstract We study the power of four query models
More informationBioinformatics - Lecture 07
Bioinformatics - Lecture 07 Bioinformatics Clusters and networks Martin Saturka http://www.bioplexity.org/lectures/ EBI version 0.4 Creative Commons Attribution-Share Alike 2.5 License Learning on profiles
More informationOne-mode Additive Clustering of Multiway Data
One-mode Additive Clustering of Multiway Data Dirk Depril and Iven Van Mechelen KULeuven Tiensestraat 103 3000 Leuven, Belgium (e-mail: dirk.depril@psy.kuleuven.ac.be iven.vanmechelen@psy.kuleuven.ac.be)
More informationEstimating Human Pose in Images. Navraj Singh December 11, 2009
Estimating Human Pose in Images Navraj Singh December 11, 2009 Introduction This project attempts to improve the performance of an existing method of estimating the pose of humans in still images. Tasks
More informationOn the Approximability of Modularity Clustering
On the Approximability of Modularity Clustering Newman s Community Finding Approach for Social Nets Bhaskar DasGupta Department of Computer Science University of Illinois at Chicago Chicago, IL 60607,
More informationOn Dimensionality Reduction of Massive Graphs for Indexing and Retrieval
On Dimensionality Reduction of Massive Graphs for Indexing and Retrieval Charu C. Aggarwal 1, Haixun Wang # IBM T. J. Watson Research Center Hawthorne, NY 153, USA 1 charu@us.ibm.com # Microsoft Research
More informationChapter 9. Classification and Clustering
Chapter 9 Classification and Clustering Classification and Clustering Classification and clustering are classical pattern recognition and machine learning problems Classification, also referred to as categorization
More informationGRAPH THEORY: AN INTRODUCTION
GRAPH THEORY: AN INTRODUCTION BEGINNERS 3/4/2018 1. GRAPHS AND THEIR PROPERTIES A graph G consists of two sets: a set of vertices V, and a set of edges E. A vertex is simply a labeled point. An edge is
More informationLatent Class Modeling as a Probabilistic Extension of K-Means Clustering
Latent Class Modeling as a Probabilistic Extension of K-Means Clustering Latent Class Cluster Models According to Kaufman and Rousseeuw (1990), cluster analysis is "the classification of similar objects
More information