Theoretical Foundations of Clustering. Margareta Ackerman

Size: px
Start display at page:

Download "Theoretical Foundations of Clustering. Margareta Ackerman"

Transcription

1 Theoretical Foundations of Clustering Margareta Ackerman

2 The Theory-Practice Gap Clustering is one of the most widely used tools for exploratory data analysis. Identifying target markets Constructing phylogenetic trees Facility allocation for city planning Personalization...

3 The Theory-Practice Gap While the interest in and application of cluster analysis has been rising rapidly, the abstract nature of the tool is still poorly understood -Wright, There has been relatively little work aimed at reasoning about clustering independently of any particular algorithm, objective function, or generative data model -Kleinberg, 2002.

4 Inherent obstacles: Clustering is ill-defined Clustering aims to organize data into groups of similar items, but beyond that There is very little consensus on the definition of clustering

5 Clustering algorithms: A few classical examples How can we partition data into k groups?

6 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage)

7 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage) Find the minimum cut (motivates spectral clustering methods)

8 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage) Find the minimum cut (motivates spectral clustering methods) Find k centers that minimize the average distance to a center (k-median, k-means,...) Many more...

9 Inherent obstacles: Clustering is inherently ambiguous There are many clustering algorithms with different (often implicit) objective functions Different algorithms have radically different input-output behavior There may be multiple reasonable clusterings There is usually no ground truth

10 Different input-output behavior of clustering algorithms

11 Different input-output behavior of clustering algorithms

12 Progress despite these obstacles: Overview Axioms of clustering quality measures (Ackerman & Ben-David, 08) Study and compare notions of clusterability (Ackerman and Ben-David, 09) Characterizing linkage-based algorithms (Ackerman, Ben-David, and Loker, 2010) Framework for clustering algorithm selection (Ackerman, Ben-David, and Loker, 2010) Characterizing hierarchical linkage-based algorithms (Ackerman & Ben- David, 2011) Properties of Phylogenetic algorithms (Ackerman, Brown, and Loker, 2012) Properties in the weighted clustering setting (Ackerman, Ben-David, Branzei, and Loker, 2012) Clustering oligarchies (Ackerman, Ben-David, Loker, and Sabato, 2013) Perturbation robust clustering (Ackerman & Schulman, 2013) Online clustering (Ackerman & Dasgupta, 2014)

13 Progress despite these obstacles: Overview Axioms of clustering quality measures (Ackerman & Ben-David, 08) Study and compare notions of clusterability (Ackerman and Ben-David, 09) Characterizing linkage-based algorithms (Ackerman, Ben-David, and Loker, 2010) Framework for clustering algorithm selection (Ackerman, Ben-David, and Loker, 2010) Characterizing hierarchical linkage-based algorithms (Ackerman & Ben- David, 2011) Properties of Phylogenetic algorithms (Ackerman, Brown, and Loker, 2012) Properties in the weighted clustering setting (Ackerman, Ben-David, Branzei, and Loker, 2012) Clustering oligarchies (Ackerman, Ben-David, Loker, and Sabato, 2013) Perturbation robust clustering (Ackerman & Schulman, 2013) Online clustering (Ackerman & Dasgupta, 2014)

14 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering

15 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering

16 Formal setup For a finite domain set X, a distance function d is the distance defined between the domain points. A clustering function maps Input: a distance function to d over X Output: a partition (clustering) of X

17 Kleinberg s axioms Scale Invariance: f(cd) =f(d) for all d and all strictly positive c. Consistency: If equals d, except for shrinking distances within clusters of f(d) or stretching between-cluster distances, then f(d 0 )=f(d). d 0 Richness: For any clustering C function over so that d X of, there exists a distance X f(d) =C.

18 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms.

19 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms. Why are axioms that seem to capture our intuition about clustering inconsistent??

20 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms. Why are axioms that seem to capture our intuition about clustering inconsistent?? Our answer: The formalization of these axioms is stronger than the intuition they intend to capture We express that same intuition in an alternative framework, and achieve consistency.

21 Clustering quality measures How good is this clustering? Clustering-quality measures quantify the quality of clusterings.

22 Defining clustering quality measures A clustering-quality measure is a function m(dataset, clustering) 2 R satisfying some properties that make this function a meaningful clustering quality measure. What properties should it satisfy?

23 Rephrasing Kleinberg s axioms for clustering quality measures Scale Invariance m(c, d) =m(c, d) for all C, d and strictly positive. Richness For any clustering C of X, there exists a distance function d over X so that C = argmax C m(c, d)

24 Consistency: If equals d, except for shrinking distances within clusters of C or stretching between-cluster distances, then m(c, d) apple m(c, d 0 ). d 0 d d 0 C C

25 Major gain - consistency of new axioms Theorem [Ackerman & Ben-David, NIPS 08]: Consistency, scale invariance, and richness for clustering quality measures form a consistent set of requirements. Dunn s index ( 73): min x6 C y d(x, y) max x C y d(x, y) This clustering quality measure satisfies consistency, scale-invariance, and richness.

26 Additional measures satisfying our axioms C-index (Dalrymple-Alford, 1970) Gamma (Baker & Hubert, 1975) Adjusted ratio of clustering (Roenker et al., 1971) D-index (Dalrymple-Alford, 1970) Modified ratio of repetition (Bower, Lesgold, and Tieman, 1969) Variations of Dunn s index (Bezdek and Pal, 1998) Strict separation (Balacan, Blum, and Vempala, 2008) And many more...

27 Why is the quality measure formulation more faithful to intuition? In the earlier setting of clustering functions, consistent changes to the underlying distance should not create any new contenders for the best clustering of the data. d C d 0 C 0 A clustering function that satisfies Kleinberg s Consistency cannot output C 0.

28 Why is the quality measure formulation more faithful to intuition? d In the setting of clustering-quality measures, consistency requires only that the quality of clustering Cnot get worse. d 0 C 0 C A different clustering can have better quality than the original.

29 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering

30 Clustering algorithm selection There is a wide variety of clustering algorithms, which can produce very different clusterings. How should a user decide which algorithm to use for a given application? 30

31 Clustering algorithm selection Users rely on cost related considerations: running times, space usage, software purchasing costs, etc There is inadequate emphasis on input-output behavior 31

32 Our framework for algorithm selection We propose a framework that lets a user utilize prior knowledge to select an algorithm Identify properties that distinguish between different input-output behavior of clustering paradigms The properties should be: 1) Intuitive and user-friendly 2) Useful for distinguishing clustering algorithms Ex. Kleinberg s axioms, order invariance, etc.. 32

33 Property-based classification for fixed k Ackerman, Ben-David, and Loker, NIPS 2010 Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 33

34 Kleinberg s axioms for fixed k Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage Kleinberg s Axioms are consistent when k is given K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 34

35 Single-linkage satisfies everything Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Recall: Single linkage is Kruskal s algorithm for Minimum Spanning Tree. It s not a good clustering algorithm in practice! 35

36 Classification in Weighted Setting Ackerman, Ben-David, Branzei, and Loker (AAAI, 2012) Weight robust: ignores element duplicates Weight sensitive: output can always be changed by duplicating some of the data Weight considering: element duplication effects the output on some data sets, but not others.

37 Classification in Weighted Setting Ackerman, Ben-David, Branzei, and Loker (AAAI, 2012) Weight robust: ignores element duplicates Weight sensitive: output can always be changed by duplicating some of the data Weight considering: element duplication effects the output on some data sets, but not others. Partitional Hierarchical Weight Robust Min Diameter k-center Single Linkage Complete Linkage Weight Sensitive k-means, k-medoids, k-median, min-sum Ward s Method Bisecting k-means Weight Considering Ratio Cut Average Linkage

38 Using property-based classification to choose an algorithm Enables users to identify a suitable algorithm without the overhead of executing many algorithms This framework helps understand the behavior of existing and new algorithms The long-term goal is to construct a property-based classification for many useful clustering algorithms 38

39 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing linkage-based clustering

40 Characterizing Linkage-Based Clustering We characterize a popular family of clustering algorithms, called linkage-based. We show that 1) all linkage-based algorithms satisfy two natural properties, and 2) no algorithm outside that family satisfies these properties.

41 Formal setting: Dendrograms and clusterings C i is a cluster in a dendrogram D if there exists a node in the dendrogram so that leaf descendants. C i is the set of its 41

42 Formal setting: Dendrograms and clusterings C = {C 1,...,C k } is a clustering in a dendrogram D if is a cluster in D for all 1 apple i apple k, and C i Clusters are disjoint 42

43 Formal setting: Hierarchical clustering algorithm A Hierarchical Clustering Algorithm A maps Input: A data set to X with a distance function d, Output: A dendrogram of X 43

44 Linkage-based algorithms Create a leaf node for every elements of X 44

45 Linkage-based algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes 45

46 Linkage-based algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes Merge the closest pair of clusters by assigning them a common parent node 46

47 Linkage-Based Algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes Merge the closest pair of clusters by assigning them a common parent node? 47

48 Examples of linkage-based algorithms The choice of linkage function distinguishes between different linkage-based algorithms. Examples of common linkage-functions Single-linkage: min between-cluster distance Average-linkage: average between-cluster distance Complete-linkage: max between-cluster distance 48

49 Characterizing Linkage-Based Clustering Partitional Setting Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 49

50 Characterizing Linkage-Based Clustering Ackerman, Ben-David, and Loker, COLT 2010 Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage The 2010 characterization applies in the partitional setting, by using the k-stopping criteria. This characterization distinguished linkage-based algorithms from other partitional techniques. 50

51 Characterizing Linkage-Based Clustering in the Hierarchal Setting (Ackerman & Ben-David, IJCAI 11) Propose two intuitive properties that uniquely identify hierarchical linkage-based clustering algorithms. Show that common hierarchical algorithms, including bisecting k-means, cannot be simulated by any linkage-based algorithm 51

52 Locality D = A(X, d) D 0 = A(X 0,d) X 0 = {x 1,...,x 4 } If we select a cluster from a dendrogram, and run the algorithm on the data in this cluster, we obtain a result that is consistent with the original dendrogram. 52

53 Outer consistency A(X,d) C (X, d) C outer-consistent change (X, d 0 ) C If A is outer-consistent, then A(X, d 0 ) will include the clustering C. 53

54 Theorem [Ackerman & Ben-David, IJCAI 11]: A hierarchical clustering algorithm is Linkage-Based if and only if it is Local and Outer-Consistent. 54

55 Easy direction of proof Every linkage-based hierarchical clustering algorithm is Local and Outer-Consistent. The proof is quite straightforward. 55

56 Interesting direction of proof If A is Local and Outer-Consistent, then A is linkage-based. To prove this direction we first need to formalize linkage-based clustering, by formally defining what is a Linkage Function. 56

57 What do we expect from a linage function? A linkage function ` : {(X 1,X 2,d):d is a distance function over X 1 [ X 2 }! R + satisfies the following: Monotonicity: If we increase distances that go between then `(X 1,X 2,d) doesn t decrease X 1 and X 2 Representation independence: Doesn t change if we re-label data X 1 X 2 57

58 Proof Sketch Recall direction: If A satisfies Outer-Consistency and Locality, then it is linkage-based. Goal Define a linkage function ` so that the linkage-based clustering based on ` outputs A(X, d) (for every X and d). 58

59 Proof Sketch Define an operator : are merged before Prove that can be extended to a partial ordering. < A (X, Y, d 1 ) < A (Z, W, d 2 ) if when we run (X [ Y [ Z [ W, d), d d 1 d 2, X Y < A on where extends and and and Z W. A A(X, d) Use the ordering to define `. Z W X Y 59

60 Sketch of proof continue: Show that < A is a partial ordering We show that < A is cycle-free. Lemma: Given a hierarchical algorithm A that is Local and Outer-Consistent, there exists no finite sequence so that (X 1,Y 1,d 1 ) < A < A (X n,y n,d n ) < A (X 1,Y 1,d 1 ). 60

61 Proof Sketch (continued...) By the above Lemma, the transitive closure of is a partial ordering. < A This implies that there exists an order preserving function ` that maps pairs of data sets to R +. It can be shown that ` satisfies the properties of a Linkage Function. 61

62 Future Directions Identify properties that are significant for specific clustering applications (some previous work in this directions by Ackerman, Brown, and Loker (ICCABS, 2012)). Analyze clustering algorithms in alternative settings, such as categorical data, fuzzy clustering, and using a noise bucket Online clustering Axiomatize clustering functions

Clustering Algorithms. Margareta Ackerman

Clustering Algorithms. Margareta Ackerman Clustering Algorithms Margareta Ackerman A sea of algorithms As we discussed last class, there are MANY clustering algorithms, and new ones are proposed all the time. They are very different from each

More information

Towards Theoretical Foundations of Clustering

Towards Theoretical Foundations of Clustering Towards Theoretical Foundations of Clustering by Margareta Ackerman A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philosophy in Computer

More information

John Oliver from The Daily Show. Supporting worthy causes at the G20 Pittsburgh Summit: Bayesians Against Discrimination. Ban Genetic Algorithms

John Oliver from The Daily Show. Supporting worthy causes at the G20 Pittsburgh Summit: Bayesians Against Discrimination. Ban Genetic Algorithms John Oliver from The Daily Show Supporting worthy causes at the G20 Pittsburgh Summit: Bayesians Against Discrimination Ban Genetic Algorithms Support Vector Machines Watch out for the protests tonight

More information

Measures of Clustering Quality: A Working Set of Axioms for Clustering

Measures of Clustering Quality: A Working Set of Axioms for Clustering Measures of Clustering Quality: A Working Set of Axioms for Clustering Margareta Ackerman and Shai Ben-David School of Computer Science University of Waterloo, Canada Abstract Aiming towards the development

More information

A Characterization of Linkage-Based Hierarchical Clustering

A Characterization of Linkage-Based Hierarchical Clustering Journal of Machine Learning Research 17 (2016) 1-17 Submitted 7/11; Revised 8/15; Published 12/16 A Characterization of Linkage-Based Hierarchical Clustering Margareta Ackerman Department of Computer Science

More information

Selecting Clustering Algorithms Based on Their Weight Sensitivity

Selecting Clustering Algorithms Based on Their Weight Sensitivity Journal of Machine Learning Research (2014) Submitted ; Published Selecting Clustering Algorithms Based on Their Weight Sensitivity Margareta Ackerman Department of Computer Science Florida State University

More information

Lecture 5 Finding meaningful clusters in data. 5.1 Kleinberg s axiomatic framework for clustering

Lecture 5 Finding meaningful clusters in data. 5.1 Kleinberg s axiomatic framework for clustering CSE 291: Unsupervised learning Spring 2008 Lecture 5 Finding meaningful clusters in data So far we ve been in the vector quantization mindset, where we want to approximate a data set by a small number

More information

Foundations of Perturbation Robust Clustering Jarrod Moore * and Margareta Ackerman * Florida State University, Tallahassee, Fl.; San José State University, San Jose, California Email: jdm0c@my.fsu.edu;

More information

Theoretical Foundations of Clustering few results, many challenges

Theoretical Foundations of Clustering few results, many challenges Theoretical Foundations of Clustering few results, many challenges Shai Ben-David University of Waterloo MLSS, Beijing, June 2014 High level view of (Statistical) Machine Learning The purpose of science

More information

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering An unsupervised machine learning problem Grouping a set of objects in such a way that objects in the same group (a cluster) are more similar (in some sense or another) to each other than to those in other

More information

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering is one of the fundamental and ubiquitous tasks in exploratory data analysis a first intuition about the

More information

Lecture 4 Hierarchical clustering

Lecture 4 Hierarchical clustering CSE : Unsupervised learning Spring 00 Lecture Hierarchical clustering. Multiple levels of granularity So far we ve talked about the k-center, k-means, and k-medoid problems, all of which involve pre-specifying

More information

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation

More information

Cluster Analysis. Ying Shen, SSE, Tongji University

Cluster Analysis. Ying Shen, SSE, Tongji University Cluster Analysis Ying Shen, SSE, Tongji University Cluster analysis Cluster analysis groups data objects based only on the attributes in the data. The main objective is that The objects within a group

More information

Hierarchical Clustering

Hierarchical Clustering Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram A tree like diagram that records the sequences of merges or splits 0 0 0 00

More information

Introduction to spectral clustering

Introduction to spectral clustering Introduction to spectral clustering Vasileios Zografos zografos@isy.liu.se Klas Nordberg klas@isy.liu.se What this course is Basic introduction into the core ideas of spectral clustering Sufficient to

More information

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Slides From Lecture Notes for Chapter 8. Introduction to Data Mining

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Slides From Lecture Notes for Chapter 8. Introduction to Data Mining Data Mining Cluster Analysis: Basic Concepts and Algorithms Slides From Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining

More information

Clustering Lecture 3: Hierarchical Methods

Clustering Lecture 3: Hierarchical Methods Clustering Lecture 3: Hierarchical Methods Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced

More information

Lecture Notes for Chapter 7. Introduction to Data Mining, 2 nd Edition. by Tan, Steinbach, Karpatne, Kumar

Lecture Notes for Chapter 7. Introduction to Data Mining, 2 nd Edition. by Tan, Steinbach, Karpatne, Kumar Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 7 Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Hierarchical Clustering Produces a set

More information

Foundations of Computer Science Spring Mathematical Preliminaries

Foundations of Computer Science Spring Mathematical Preliminaries Foundations of Computer Science Spring 2017 Equivalence Relation, Recursive Definition, and Mathematical Induction Mathematical Preliminaries Mohammad Ashiqur Rahman Department of Computer Science College

More information

Optimization I : Brute force and Greedy strategy

Optimization I : Brute force and Greedy strategy Chapter 3 Optimization I : Brute force and Greedy strategy A generic definition of an optimization problem involves a set of constraints that defines a subset in some underlying space (like the Euclidean

More information

Hierarchical Clustering: Objectives & Algorithms. École normale supérieure & CNRS

Hierarchical Clustering: Objectives & Algorithms. École normale supérieure & CNRS Hierarchical Clustering: Objectives & Algorithms Vincent Cohen-Addad Paris Sorbonne & CNRS Frederik Mallmann-Trenn MIT Varun Kanade University of Oxford Claire Mathieu École normale supérieure & CNRS Clustering

More information

Generalized Transitive Distance with Minimum Spanning Random Forest

Generalized Transitive Distance with Minimum Spanning Random Forest Generalized Transitive Distance with Minimum Spanning Random Forest Author: Zhiding Yu and B. V. K. Vijaya Kumar, et al. Dept of Electrical and Computer Engineering Carnegie Mellon University 1 Clustering

More information

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Cluster Analysis Mu-Chun Su Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Introduction Cluster analysis is the formal study of algorithms and methods

More information

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search Informal goal Clustering Given set of objects and measure of similarity between them, group similar objects together What mean by similar? What is good grouping? Computation time / quality tradeoff 1 2

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

Data Mining Concepts & Techniques

Data Mining Concepts & Techniques Data Mining Concepts & Techniques Lecture No 08 Cluster Analysis Naeem Ahmed Email: naeemmahoto@gmailcom Department of Software Engineering Mehran Univeristy of Engineering and Technology Jamshoro Outline

More information

An Axiomatic Approach to Routing. Omer Lev, Moshe Tennenholtz and Aviv Zohar TARK 2015

An Axiomatic Approach to Routing. Omer Lev, Moshe Tennenholtz and Aviv Zohar TARK 2015 An Axiomatic Approach to Routing Omer Lev, Moshe Tennenholtz and Aviv Zohar TARK 205 What is routing? A network of connected nodes with edge weights 5 3 2 3 4 What is routing? A network of connected nodes

More information

CS7267 MACHINE LEARNING

CS7267 MACHINE LEARNING S7267 MAHINE LEARNING HIERARHIAL LUSTERING Ref: hengkai Li, Department of omputer Science and Engineering, University of Texas at Arlington (Slides courtesy of Vipin Kumar) Mingon Kang, Ph.D. omputer Science,

More information

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06 Clustering CS294 Practical Machine Learning Junming Yin 10/09/06 Outline Introduction Unsupervised learning What is clustering? Application Dissimilarity (similarity) of objects Clustering algorithm K-means,

More information

A Computational Theory of Clustering

A Computational Theory of Clustering A Computational Theory of Clustering Avrim Blum Carnegie Mellon University Based on work joint with Nina Balcan, Anupam Gupta, and Santosh Vempala Point of this talk A new way to theoretically analyze

More information

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler BBS654 Data Mining Pinar Duygulu Slides are adapted from Nazli Ikizler 1 Classification Classification systems: Supervised learning Make a rational prediction given evidence There are several methods for

More information

CS264: Beyond Worst-Case Analysis Lecture #6: Perturbation-Stable Clustering

CS264: Beyond Worst-Case Analysis Lecture #6: Perturbation-Stable Clustering CS264: Beyond Worst-Case Analysis Lecture #6: Perturbation-Stable Clustering Tim Roughgarden January 26, 2017 1 Clustering Is Hard Only When It Doesn t Matter In some optimization problems, the objective

More information

Chapter 6: Cluster Analysis

Chapter 6: Cluster Analysis Chapter 6: Cluster Analysis The major goal of cluster analysis is to separate individual observations, or items, into groups, or clusters, on the basis of the values for the q variables measured on each

More information

Hierarchical Graph Clustering: Quality Metrics & Algorithms

Hierarchical Graph Clustering: Quality Metrics & Algorithms Hierarchical Graph Clustering: Quality Metrics & Algorithms Thomas Bonald Joint work with Bertrand Charpentier, Alexis Galland & Alexandre Hollocou LTCI Data Science seminar March 2019 Motivation Clustering

More information

Hierarchical Clustering

Hierarchical Clustering What is clustering Partitioning of a data set into subsets. A cluster is a group of relatively homogeneous cases or observations Hierarchical Clustering Mikhail Dozmorov Fall 2016 2/61 What is clustering

More information

Clustering Part 3. Hierarchical Clustering

Clustering Part 3. Hierarchical Clustering Clustering Part Dr Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville Hierarchical Clustering Two main types: Agglomerative Start with the points

More information

Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees

Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees Lior Kamma 1 Introduction Embeddings and Distortion An embedding of a metric space (X, d X ) into a metric space (Y, d Y ) is

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

CS 534: Computer Vision Segmentation and Perceptual Grouping

CS 534: Computer Vision Segmentation and Perceptual Grouping CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation

More information

Minimum Spanning Trees

Minimum Spanning Trees Minimum Spanning Trees Overview Problem A town has a set of houses and a set of roads. A road connects and only houses. A road connecting houses u and v has a repair cost w(u, v). Goal: Repair enough (and

More information

Clustering Algorithms for general similarity measures

Clustering Algorithms for general similarity measures Types of general clustering methods Clustering Algorithms for general similarity measures general similarity measure: specified by object X object similarity matrix 1 constructive algorithms agglomerative

More information

Clustering algorithms and introduction to persistent homology

Clustering algorithms and introduction to persistent homology Foundations of Geometric Methods in Data Analysis 2017-18 Clustering algorithms and introduction to persistent homology Frédéric Chazal INRIA Saclay - Ile-de-France frederic.chazal@inria.fr Introduction

More information

Scalable Clustering of Signed Networks Using Balance Normalized Cut

Scalable Clustering of Signed Networks Using Balance Normalized Cut Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.

More information

Representation Learning for Clustering: A Statistical Framework

Representation Learning for Clustering: A Statistical Framework Representation Learning for Clustering: A Statistical Framework Hassan Ashtiani School of Computer Science University of Waterloo mhzokaei@uwaterloo.ca Shai Ben-David School of Computer Science University

More information

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010 Cluster Analysis Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX 7575 April 008 April 010 Cluster Analysis, sometimes called data segmentation or customer segmentation,

More information

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs

Graphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs Graphs and Network Flows IE411 Lecture 21 Dr. Ted Ralphs IE411 Lecture 21 1 Combinatorial Optimization and Network Flows In general, most combinatorial optimization and integer programming problems are

More information

BMI/STAT 768: Lecture 06 Trees in Graphs

BMI/STAT 768: Lecture 06 Trees in Graphs BMI/STAT 768: Lecture 06 Trees in Graphs Moo K. Chung mkchung@wisc.edu February 11, 2018 Parts of this lecture is based on [3, 5]. Many objects and data can be represented as networks. Unfortunately networks

More information

10. Clustering. Introduction to Bioinformatics Jarkko Salojärvi. Based on lecture slides by Samuel Kaski

10. Clustering. Introduction to Bioinformatics Jarkko Salojärvi. Based on lecture slides by Samuel Kaski 10. Clustering Introduction to Bioinformatics 30.9.2008 Jarkko Salojärvi Based on lecture slides by Samuel Kaski Definition of a cluster Typically either 1. A group of mutually similar samples, or 2. A

More information

Lecture 5: Duality Theory

Lecture 5: Duality Theory Lecture 5: Duality Theory Rajat Mittal IIT Kanpur The objective of this lecture note will be to learn duality theory of linear programming. We are planning to answer following questions. What are hyperplane

More information

4. Ad-hoc I: Hierarchical clustering

4. Ad-hoc I: Hierarchical clustering 4. Ad-hoc I: Hierarchical clustering Hierarchical versus Flat Flat methods generate a single partition into k clusters. The number k of clusters has to be determined by the user ahead of time. Hierarchical

More information

STATS306B STATS306B. Clustering. Jonathan Taylor Department of Statistics Stanford University. June 3, 2010

STATS306B STATS306B. Clustering. Jonathan Taylor Department of Statistics Stanford University. June 3, 2010 STATS306B Jonathan Taylor Department of Statistics Stanford University June 3, 2010 Spring 2010 Outline K-means, K-medoids, EM algorithm choosing number of clusters: Gap test hierarchical clustering spectral

More information

10701 Machine Learning. Clustering

10701 Machine Learning. Clustering 171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among

More information

Hierarchical Clustering

Hierarchical Clustering Hierarchical Clustering Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram A tree-like diagram that records the sequences of merges

More information

On the Relationships between Zero Forcing Numbers and Certain Graph Coverings

On the Relationships between Zero Forcing Numbers and Certain Graph Coverings On the Relationships between Zero Forcing Numbers and Certain Graph Coverings Fatemeh Alinaghipour Taklimi, Shaun Fallat 1,, Karen Meagher 2 Department of Mathematics and Statistics, University of Regina,

More information

The Encoding Complexity of Network Coding

The Encoding Complexity of Network Coding The Encoding Complexity of Network Coding Michael Langberg Alexander Sprintson Jehoshua Bruck California Institute of Technology Email: mikel,spalex,bruck @caltech.edu Abstract In the multicast network

More information

Matching Algorithms. Proof. If a bipartite graph has a perfect matching, then it is easy to see that the right hand side is a necessary condition.

Matching Algorithms. Proof. If a bipartite graph has a perfect matching, then it is easy to see that the right hand side is a necessary condition. 18.433 Combinatorial Optimization Matching Algorithms September 9,14,16 Lecturer: Santosh Vempala Given a graph G = (V, E), a matching M is a set of edges with the property that no two of the edges have

More information

What to come. There will be a few more topics we will cover on supervised learning

What to come. There will be a few more topics we will cover on supervised learning Summary so far Supervised learning learn to predict Continuous target regression; Categorical target classification Linear Regression Classification Discriminative models Perceptron (linear) Logistic regression

More information

Section 3 Default Logic. Subsection 3.1 Introducing defaults and default logics

Section 3 Default Logic. Subsection 3.1 Introducing defaults and default logics Section 3 Default Logic Subsection 3.1 Introducing defaults and default logics TU Dresden, WS 2017/18 Introduction to Nonmonotonic Reasoning Slide 34 Introducing defaults and default logics: an example

More information

Cluster Analysis for Microarray Data

Cluster Analysis for Microarray Data Cluster Analysis for Microarray Data Seventh International Long Oligonucleotide Microarray Workshop Tucson, Arizona January 7-12, 2007 Dan Nettleton IOWA STATE UNIVERSITY 1 Clustering Group objects that

More information

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006

2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 The Encoding Complexity of Network Coding Michael Langberg, Member, IEEE, Alexander Sprintson, Member, IEEE, and Jehoshua Bruck,

More information

Reading 1 : Introduction

Reading 1 : Introduction CS/Math 240: Introduction to Discrete Mathematics Fall 2015 Instructors: Beck Hasti and Gautam Prakriya Reading 1 : Introduction Welcome to CS 240, an introduction to discrete mathematics. This reading

More information

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm DATA MINING LECTURE 7 Hierarchical Clustering, DBSCAN The EM Algorithm CLUSTERING What is a Clustering? In general a grouping of objects such that the objects in a group (cluster) are similar (or related)

More information

Types of general clustering methods. Clustering Algorithms for general similarity measures. Similarity between clusters

Types of general clustering methods. Clustering Algorithms for general similarity measures. Similarity between clusters Types of general clustering methods Clustering Algorithms for general similarity measures agglomerative versus divisive algorithms agglomerative = bottom-up build up clusters from single objects divisive

More information

Clustering. Chapter 10 in Introduction to statistical learning

Clustering. Chapter 10 in Introduction to statistical learning Clustering Chapter 10 in Introduction to statistical learning 16 14 12 10 8 6 4 2 0 2 4 6 8 10 12 14 1 Clustering ² Clustering is the art of finding groups in data (Kaufman and Rousseeuw, 1990). ² What

More information

ECS 234: Data Analysis: Clustering ECS 234

ECS 234: Data Analysis: Clustering ECS 234 : Data Analysis: Clustering What is Clustering? Given n objects, assign them to groups (clusters) based on their similarity Unsupervised Machine Learning Class Discovery Difficult, and maybe ill-posed

More information

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining 4/18/004 1

More information

Unsupervised Learning and Clustering

Unsupervised Learning and Clustering Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2008 CS 551, Spring 2008 c 2008, Selim Aksoy (Bilkent University)

More information

Lesson 3. Prof. Enza Messina

Lesson 3. Prof. Enza Messina Lesson 3 Prof. Enza Messina Clustering techniques are generally classified into these classes: PARTITIONING ALGORITHMS Directly divides data points into some prespecified number of clusters without a hierarchical

More information

A synchronizer generates sequences of clock pulses at each node of the network satisfying the condition given by the following definition.

A synchronizer generates sequences of clock pulses at each node of the network satisfying the condition given by the following definition. Chapter 8 Synchronizers So far, we have mainly studied synchronous algorithms because generally, asynchronous algorithms are often more di cult to obtain and it is substantially harder to reason about

More information

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22 INF4820 Clustering Erik Velldal University of Oslo Nov. 17, 2009 Erik Velldal INF4820 1 / 22 Topics for Today More on unsupervised machine learning for data-driven categorization: clustering. The task

More information

CSE 5243 INTRO. TO DATA MINING

CSE 5243 INTRO. TO DATA MINING CSE 5243 INTRO. TO DATA MINING Cluster Analysis: Basic Concepts and Methods Huan Sun, CSE@The Ohio State University Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han 2 Chapter 10. Cluster

More information

Representing Product Designs Using a Description Graph Extension to OWL 2

Representing Product Designs Using a Description Graph Extension to OWL 2 Representing Product Designs Using a Description Graph Extension to OWL 2 Henson Graves Lockheed Martin Aeronautics Company Fort Worth Texas, USA henson.graves@lmco.com Abstract. Product development requires

More information

Clustering. (Part 2)

Clustering. (Part 2) Clustering (Part 2) 1 k-means clustering 2 General Observations on k-means clustering In essence, k-means clustering aims at minimizing cluster variance. It is typically used in Euclidean spaces and works

More information

Thoughts on clustering

Thoughts on clustering Thoughts on clustering Avrim Blum Carnegie Mellon University Abstract Clustering is a somewhat confusing topic theoretically. In large part this is because there are many different kinds of clustering

More information

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017)

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017) 1 Notes Reminder: HW2 Due Today by 11:59PM TA s note: Please provide a detailed ReadMe.txt file on how to run the program on the STDLINUX. If you installed/upgraded any package on STDLINUX, you should

More information

Data Clustering. Danushka Bollegala

Data Clustering. Danushka Bollegala Data Clustering Danushka Bollegala Outline Why cluster data? Clustering as unsupervised learning Clustering algorithms k-means, k-medoids agglomerative clustering Brown s clustering Spectral clustering

More information

On Constraint Problems with Incomplete or Erroneous Data

On Constraint Problems with Incomplete or Erroneous Data On Constraint Problems with Incomplete or Erroneous Data Neil Yorke-Smith and Carmen Gervet IC Parc, Imperial College, London, SW7 2AZ, U.K. nys,cg6 @icparc.ic.ac.uk Abstract. Real-world constraint problems

More information

Algorithms, Games, and Networks February 21, Lecture 12

Algorithms, Games, and Networks February 21, Lecture 12 Algorithms, Games, and Networks February, 03 Lecturer: Ariel Procaccia Lecture Scribe: Sercan Yıldız Overview In this lecture, we introduce the axiomatic approach to social choice theory. In particular,

More information

Hierarchical clustering

Hierarchical clustering Hierarchical clustering Based in part on slides from textbook, slides of Susan Holmes December 2, 2012 1 / 1 Description Produces a set of nested clusters organized as a hierarchical tree. Can be visualized

More information

Cluster analysis. Agnieszka Nowak - Brzezinska

Cluster analysis. Agnieszka Nowak - Brzezinska Cluster analysis Agnieszka Nowak - Brzezinska Outline of lecture What is cluster analysis? Clustering algorithms Measures of Cluster Validity What is Cluster Analysis? Finding groups of objects such that

More information

Cluster Analysis. Angela Montanari and Laura Anderlucci

Cluster Analysis. Angela Montanari and Laura Anderlucci Cluster Analysis Angela Montanari and Laura Anderlucci 1 Introduction Clustering a set of n objects into k groups is usually moved by the aim of identifying internally homogenous groups according to a

More information

MSA220 - Statistical Learning for Big Data

MSA220 - Statistical Learning for Big Data MSA220 - Statistical Learning for Big Data Lecture 13 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Clustering Explorative analysis - finding groups

More information

UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 8 Lecturer: David Wagner February 20, Notes 8 for CS 170

UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 8 Lecturer: David Wagner February 20, Notes 8 for CS 170 UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 8 Lecturer: David Wagner February 20, 2003 Notes 8 for CS 170 1 Minimum Spanning Trees A tree is an undirected graph that is connected

More information

CS570: Introduction to Data Mining

CS570: Introduction to Data Mining CS570: Introduction to Data Mining Scalable Clustering Methods: BIRCH and Others Reading: Chapter 10.3 Han, Chapter 9.5 Tan Cengiz Gunay, Ph.D. Slides courtesy of Li Xiong, Ph.D., 2011 Han, Kamber & Pei.

More information

Theorem 2.9: nearest addition algorithm

Theorem 2.9: nearest addition algorithm There are severe limits on our ability to compute near-optimal tours It is NP-complete to decide whether a given undirected =(,)has a Hamiltonian cycle An approximation algorithm for the TSP can be used

More information

Throughout the chapter, we will assume that the reader is familiar with the basics of phylogenetic trees.

Throughout the chapter, we will assume that the reader is familiar with the basics of phylogenetic trees. Chapter 7 SUPERTREE ALGORITHMS FOR NESTED TAXA Philip Daniel and Charles Semple Abstract: Keywords: Most supertree algorithms combine collections of rooted phylogenetic trees with overlapping leaf sets

More information

Discrete mathematics

Discrete mathematics Discrete mathematics Petr Kovář petr.kovar@vsb.cz VŠB Technical University of Ostrava DiM 470-2301/02, Winter term 2018/2019 About this file This file is meant to be a guideline for the lecturer. Many

More information

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1

Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanford.edu) January 11, 2018 Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1 In this lecture

More information

Greedy Algorithms Part Three

Greedy Algorithms Part Three Greedy Algorithms Part Three Announcements Problem Set Four due right now. Due on Wednesday with a late day. Problem Set Five out, due Monday, August 5. Explore greedy algorithms, exchange arguments, greedy

More information

Inductive datatypes in HOL. lessons learned in Formal-Logic Engineering

Inductive datatypes in HOL. lessons learned in Formal-Logic Engineering Inductive datatypes in HOL lessons learned in Formal-Logic Engineering Stefan Berghofer and Markus Wenzel Institut für Informatik TU München = Isabelle λ β HOL α 1 Introduction Applications of inductive

More information

On graphs with disjoint dominating and 2-dominating sets

On graphs with disjoint dominating and 2-dominating sets On graphs with disjoint dominating and 2-dominating sets 1 Michael A. Henning and 2 Douglas F. Rall 1 Department of Mathematics University of Johannesburg Auckland Park, 2006 South Africa Email: mahenning@uj.ac.za

More information

Tree Models of Similarity and Association. Clustering and Classification Lecture 5

Tree Models of Similarity and Association. Clustering and Classification Lecture 5 Tree Models of Similarity and Association Clustering and Lecture 5 Today s Class Tree models. Hierarchical clustering methods. Fun with ultrametrics. 2 Preliminaries Today s lecture is based on the monograph

More information

Collection Guiding: Multimedia Collection Browsing and Visualization. Outline. Context. Searching for data

Collection Guiding: Multimedia Collection Browsing and Visualization. Outline. Context. Searching for data Collection Guiding: Multimedia Collection Browsing and Visualization Stéphane Marchand-Maillet Viper CVML University of Geneva marchand@cui.unige.ch http://viper.unige.ch Outline Multimedia data context

More information

Motivation. Technical Background

Motivation. Technical Background Handling Outliers through Agglomerative Clustering with Full Model Maximum Likelihood Estimation, with Application to Flow Cytometry Mark Gordon, Justin Li, Kevin Matzen, Bryce Wiedenbeck Motivation Clustering

More information

IT-Dendrogram: A New Member of the In-Tree (IT) Clustering Family

IT-Dendrogram: A New Member of the In-Tree (IT) Clustering Family IT-Dendrogram: A New Member of the In-Tree (IT) Clustering Family Teng Qiu (qiutengcool@163.com) Yongjie Li (liyj@uestc.edu.cn) University of Electronic Science and Technology of China, Chengdu, China

More information

Ma/CS 6b Class 7: Minors

Ma/CS 6b Class 7: Minors Ma/CS 6b Class 7: Minors By Adam Sheffer Edge Subdivision iven a graph = V, E, and an edge e E, subdividing e is the operation of replacing e with a path consisting of new vertices. 1 raph Relations We

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering K-means and Hierarchical Clustering Xiaohui Xie University of California, Irvine K-means and Hierarchical Clustering p.1/18 Clustering Given n data points X = {x 1, x 2,, x n }. Clustering is the partitioning

More information

Trees. 3. (Minimally Connected) G is connected and deleting any of its edges gives rise to a disconnected graph.

Trees. 3. (Minimally Connected) G is connected and deleting any of its edges gives rise to a disconnected graph. Trees 1 Introduction Trees are very special kind of (undirected) graphs. Formally speaking, a tree is a connected graph that is acyclic. 1 This definition has some drawbacks: given a graph it is not trivial

More information