Theoretical Foundations of Clustering. Margareta Ackerman
|
|
- Stuart Powell
- 6 years ago
- Views:
Transcription
1 Theoretical Foundations of Clustering Margareta Ackerman
2 The Theory-Practice Gap Clustering is one of the most widely used tools for exploratory data analysis. Identifying target markets Constructing phylogenetic trees Facility allocation for city planning Personalization...
3 The Theory-Practice Gap While the interest in and application of cluster analysis has been rising rapidly, the abstract nature of the tool is still poorly understood -Wright, There has been relatively little work aimed at reasoning about clustering independently of any particular algorithm, objective function, or generative data model -Kleinberg, 2002.
4 Inherent obstacles: Clustering is ill-defined Clustering aims to organize data into groups of similar items, but beyond that There is very little consensus on the definition of clustering
5 Clustering algorithms: A few classical examples How can we partition data into k groups?
6 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage)
7 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage) Find the minimum cut (motivates spectral clustering methods)
8 Clustering algorithms: A few classical examples How can we partition data into k groups? Use Kruskal s algorithm for MST (Singlelinkage) Find the minimum cut (motivates spectral clustering methods) Find k centers that minimize the average distance to a center (k-median, k-means,...) Many more...
9 Inherent obstacles: Clustering is inherently ambiguous There are many clustering algorithms with different (often implicit) objective functions Different algorithms have radically different input-output behavior There may be multiple reasonable clusterings There is usually no ground truth
10 Different input-output behavior of clustering algorithms
11 Different input-output behavior of clustering algorithms
12 Progress despite these obstacles: Overview Axioms of clustering quality measures (Ackerman & Ben-David, 08) Study and compare notions of clusterability (Ackerman and Ben-David, 09) Characterizing linkage-based algorithms (Ackerman, Ben-David, and Loker, 2010) Framework for clustering algorithm selection (Ackerman, Ben-David, and Loker, 2010) Characterizing hierarchical linkage-based algorithms (Ackerman & Ben- David, 2011) Properties of Phylogenetic algorithms (Ackerman, Brown, and Loker, 2012) Properties in the weighted clustering setting (Ackerman, Ben-David, Branzei, and Loker, 2012) Clustering oligarchies (Ackerman, Ben-David, Loker, and Sabato, 2013) Perturbation robust clustering (Ackerman & Schulman, 2013) Online clustering (Ackerman & Dasgupta, 2014)
13 Progress despite these obstacles: Overview Axioms of clustering quality measures (Ackerman & Ben-David, 08) Study and compare notions of clusterability (Ackerman and Ben-David, 09) Characterizing linkage-based algorithms (Ackerman, Ben-David, and Loker, 2010) Framework for clustering algorithm selection (Ackerman, Ben-David, and Loker, 2010) Characterizing hierarchical linkage-based algorithms (Ackerman & Ben- David, 2011) Properties of Phylogenetic algorithms (Ackerman, Brown, and Loker, 2012) Properties in the weighted clustering setting (Ackerman, Ben-David, Branzei, and Loker, 2012) Clustering oligarchies (Ackerman, Ben-David, Loker, and Sabato, 2013) Perturbation robust clustering (Ackerman & Schulman, 2013) Online clustering (Ackerman & Dasgupta, 2014)
14 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering
15 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering
16 Formal setup For a finite domain set X, a distance function d is the distance defined between the domain points. A clustering function maps Input: a distance function to d over X Output: a partition (clustering) of X
17 Kleinberg s axioms Scale Invariance: f(cd) =f(d) for all d and all strictly positive c. Consistency: If equals d, except for shrinking distances within clusters of f(d) or stretching between-cluster distances, then f(d 0 )=f(d). d 0 Richness: For any clustering C function over so that d X of, there exists a distance X f(d) =C.
18 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms.
19 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms. Why are axioms that seem to capture our intuition about clustering inconsistent??
20 Theorem [Kleinberg, 02]: These axioms are inconsistent. Namely, no function can satisfy these three axioms. Why are axioms that seem to capture our intuition about clustering inconsistent?? Our answer: The formalization of these axioms is stronger than the intuition they intend to capture We express that same intuition in an alternative framework, and achieve consistency.
21 Clustering quality measures How good is this clustering? Clustering-quality measures quantify the quality of clusterings.
22 Defining clustering quality measures A clustering-quality measure is a function m(dataset, clustering) 2 R satisfying some properties that make this function a meaningful clustering quality measure. What properties should it satisfy?
23 Rephrasing Kleinberg s axioms for clustering quality measures Scale Invariance m(c, d) =m(c, d) for all C, d and strictly positive. Richness For any clustering C of X, there exists a distance function d over X so that C = argmax C m(c, d)
24 Consistency: If equals d, except for shrinking distances within clusters of C or stretching between-cluster distances, then m(c, d) apple m(c, d 0 ). d 0 d d 0 C C
25 Major gain - consistency of new axioms Theorem [Ackerman & Ben-David, NIPS 08]: Consistency, scale invariance, and richness for clustering quality measures form a consistent set of requirements. Dunn s index ( 73): min x6 C y d(x, y) max x C y d(x, y) This clustering quality measure satisfies consistency, scale-invariance, and richness.
26 Additional measures satisfying our axioms C-index (Dalrymple-Alford, 1970) Gamma (Baker & Hubert, 1975) Adjusted ratio of clustering (Roenker et al., 1971) D-index (Dalrymple-Alford, 1970) Modified ratio of repetition (Bower, Lesgold, and Tieman, 1969) Variations of Dunn s index (Bezdek and Pal, 1998) Strict separation (Balacan, Blum, and Vempala, 2008) And many more...
27 Why is the quality measure formulation more faithful to intuition? In the earlier setting of clustering functions, consistent changes to the underlying distance should not create any new contenders for the best clustering of the data. d C d 0 C 0 A clustering function that satisfies Kleinberg s Consistency cannot output C 0.
28 Why is the quality measure formulation more faithful to intuition? d In the setting of clustering-quality measures, consistency requires only that the quality of clustering Cnot get worse. d 0 C 0 C A different clustering can have better quality than the original.
29 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing Linkage-Based clustering
30 Clustering algorithm selection There is a wide variety of clustering algorithms, which can produce very different clusterings. How should a user decide which algorithm to use for a given application? 30
31 Clustering algorithm selection Users rely on cost related considerations: running times, space usage, software purchasing costs, etc There is inadequate emphasis on input-output behavior 31
32 Our framework for algorithm selection We propose a framework that lets a user utilize prior knowledge to select an algorithm Identify properties that distinguish between different input-output behavior of clustering paradigms The properties should be: 1) Intuitive and user-friendly 2) Useful for distinguishing clustering algorithms Ex. Kleinberg s axioms, order invariance, etc.. 32
33 Property-based classification for fixed k Ackerman, Ben-David, and Loker, NIPS 2010 Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 33
34 Kleinberg s axioms for fixed k Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage Kleinberg s Axioms are consistent when k is given K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 34
35 Single-linkage satisfies everything Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Recall: Single linkage is Kruskal s algorithm for Minimum Spanning Tree. It s not a good clustering algorithm in practice! 35
36 Classification in Weighted Setting Ackerman, Ben-David, Branzei, and Loker (AAAI, 2012) Weight robust: ignores element duplicates Weight sensitive: output can always be changed by duplicating some of the data Weight considering: element duplication effects the output on some data sets, but not others.
37 Classification in Weighted Setting Ackerman, Ben-David, Branzei, and Loker (AAAI, 2012) Weight robust: ignores element duplicates Weight sensitive: output can always be changed by duplicating some of the data Weight considering: element duplication effects the output on some data sets, but not others. Partitional Hierarchical Weight Robust Min Diameter k-center Single Linkage Complete Linkage Weight Sensitive k-means, k-medoids, k-median, min-sum Ward s Method Bisecting k-means Weight Considering Ratio Cut Average Linkage
38 Using property-based classification to choose an algorithm Enables users to identify a suitable algorithm without the overhead of executing many algorithms This framework helps understand the behavior of existing and new algorithms The long-term goal is to construct a property-based classification for many useful clustering algorithms 38
39 Outline Axiomatic treatment of clustering Clustering algorithm selection Characterizing linkage-based clustering
40 Characterizing Linkage-Based Clustering We characterize a popular family of clustering algorithms, called linkage-based. We show that 1) all linkage-based algorithms satisfy two natural properties, and 2) no algorithm outside that family satisfies these properties.
41 Formal setting: Dendrograms and clusterings C i is a cluster in a dendrogram D if there exists a node in the dendrogram so that leaf descendants. C i is the set of its 41
42 Formal setting: Dendrograms and clusterings C = {C 1,...,C k } is a clustering in a dendrogram D if is a cluster in D for all 1 apple i apple k, and C i Clusters are disjoint 42
43 Formal setting: Hierarchical clustering algorithm A Hierarchical Clustering Algorithm A maps Input: A data set to X with a distance function d, Output: A dendrogram of X 43
44 Linkage-based algorithms Create a leaf node for every elements of X 44
45 Linkage-based algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes 45
46 Linkage-based algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes Merge the closest pair of clusters by assigning them a common parent node 46
47 Linkage-Based Algorithms Create a leaf node for every elements of X Repeat the following until a single tree remains: Consider clusters represented by the remaining root nodes Merge the closest pair of clusters by assigning them a common parent node? 47
48 Examples of linkage-based algorithms The choice of linkage function distinguishes between different linkage-based algorithms. Examples of common linkage-functions Single-linkage: min between-cluster distance Average-linkage: average between-cluster distance Complete-linkage: max between-cluster distance 48
49 Characterizing Linkage-Based Clustering Partitional Setting Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage K-means!! " " " "!!!! K-medoids!! " " " "!!!! Min-Sum!!!! " "!!!! Ratio-cut " "!! " "!!!! Normalizedcut " " " " " "!!!! 49
50 Characterizing Linkage-Based Clustering Ackerman, Ben-David, and Loker, COLT 2010 Local Outer Con. Inner Con. Consistent Refin. Preserv Order Inv. Rich Outer Rich Rep Ind Scale Inv Single!!!!!!!!!! linkage Average!! " "! "!!!! linkage Complete!! " "!!!!!! linkage The 2010 characterization applies in the partitional setting, by using the k-stopping criteria. This characterization distinguished linkage-based algorithms from other partitional techniques. 50
51 Characterizing Linkage-Based Clustering in the Hierarchal Setting (Ackerman & Ben-David, IJCAI 11) Propose two intuitive properties that uniquely identify hierarchical linkage-based clustering algorithms. Show that common hierarchical algorithms, including bisecting k-means, cannot be simulated by any linkage-based algorithm 51
52 Locality D = A(X, d) D 0 = A(X 0,d) X 0 = {x 1,...,x 4 } If we select a cluster from a dendrogram, and run the algorithm on the data in this cluster, we obtain a result that is consistent with the original dendrogram. 52
53 Outer consistency A(X,d) C (X, d) C outer-consistent change (X, d 0 ) C If A is outer-consistent, then A(X, d 0 ) will include the clustering C. 53
54 Theorem [Ackerman & Ben-David, IJCAI 11]: A hierarchical clustering algorithm is Linkage-Based if and only if it is Local and Outer-Consistent. 54
55 Easy direction of proof Every linkage-based hierarchical clustering algorithm is Local and Outer-Consistent. The proof is quite straightforward. 55
56 Interesting direction of proof If A is Local and Outer-Consistent, then A is linkage-based. To prove this direction we first need to formalize linkage-based clustering, by formally defining what is a Linkage Function. 56
57 What do we expect from a linage function? A linkage function ` : {(X 1,X 2,d):d is a distance function over X 1 [ X 2 }! R + satisfies the following: Monotonicity: If we increase distances that go between then `(X 1,X 2,d) doesn t decrease X 1 and X 2 Representation independence: Doesn t change if we re-label data X 1 X 2 57
58 Proof Sketch Recall direction: If A satisfies Outer-Consistency and Locality, then it is linkage-based. Goal Define a linkage function ` so that the linkage-based clustering based on ` outputs A(X, d) (for every X and d). 58
59 Proof Sketch Define an operator : are merged before Prove that can be extended to a partial ordering. < A (X, Y, d 1 ) < A (Z, W, d 2 ) if when we run (X [ Y [ Z [ W, d), d d 1 d 2, X Y < A on where extends and and and Z W. A A(X, d) Use the ordering to define `. Z W X Y 59
60 Sketch of proof continue: Show that < A is a partial ordering We show that < A is cycle-free. Lemma: Given a hierarchical algorithm A that is Local and Outer-Consistent, there exists no finite sequence so that (X 1,Y 1,d 1 ) < A < A (X n,y n,d n ) < A (X 1,Y 1,d 1 ). 60
61 Proof Sketch (continued...) By the above Lemma, the transitive closure of is a partial ordering. < A This implies that there exists an order preserving function ` that maps pairs of data sets to R +. It can be shown that ` satisfies the properties of a Linkage Function. 61
62 Future Directions Identify properties that are significant for specific clustering applications (some previous work in this directions by Ackerman, Brown, and Loker (ICCABS, 2012)). Analyze clustering algorithms in alternative settings, such as categorical data, fuzzy clustering, and using a noise bucket Online clustering Axiomatize clustering functions
Clustering Algorithms. Margareta Ackerman
Clustering Algorithms Margareta Ackerman A sea of algorithms As we discussed last class, there are MANY clustering algorithms, and new ones are proposed all the time. They are very different from each
More informationTowards Theoretical Foundations of Clustering
Towards Theoretical Foundations of Clustering by Margareta Ackerman A thesis presented to the University of Waterloo in fulfillment of the thesis requirement for the degree of Doctor of Philosophy in Computer
More informationJohn Oliver from The Daily Show. Supporting worthy causes at the G20 Pittsburgh Summit: Bayesians Against Discrimination. Ban Genetic Algorithms
John Oliver from The Daily Show Supporting worthy causes at the G20 Pittsburgh Summit: Bayesians Against Discrimination Ban Genetic Algorithms Support Vector Machines Watch out for the protests tonight
More informationMeasures of Clustering Quality: A Working Set of Axioms for Clustering
Measures of Clustering Quality: A Working Set of Axioms for Clustering Margareta Ackerman and Shai Ben-David School of Computer Science University of Waterloo, Canada Abstract Aiming towards the development
More informationA Characterization of Linkage-Based Hierarchical Clustering
Journal of Machine Learning Research 17 (2016) 1-17 Submitted 7/11; Revised 8/15; Published 12/16 A Characterization of Linkage-Based Hierarchical Clustering Margareta Ackerman Department of Computer Science
More informationSelecting Clustering Algorithms Based on Their Weight Sensitivity
Journal of Machine Learning Research (2014) Submitted ; Published Selecting Clustering Algorithms Based on Their Weight Sensitivity Margareta Ackerman Department of Computer Science Florida State University
More informationLecture 5 Finding meaningful clusters in data. 5.1 Kleinberg s axiomatic framework for clustering
CSE 291: Unsupervised learning Spring 2008 Lecture 5 Finding meaningful clusters in data So far we ve been in the vector quantization mindset, where we want to approximate a data set by a small number
More informationFoundations of Perturbation Robust Clustering Jarrod Moore * and Margareta Ackerman * Florida State University, Tallahassee, Fl.; San José State University, San Jose, California Email: jdm0c@my.fsu.edu;
More informationTheoretical Foundations of Clustering few results, many challenges
Theoretical Foundations of Clustering few results, many challenges Shai Ben-David University of Waterloo MLSS, Beijing, June 2014 High level view of (Statistical) Machine Learning The purpose of science
More informationHard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering
An unsupervised machine learning problem Grouping a set of objects in such a way that objects in the same group (a cluster) are more similar (in some sense or another) to each other than to those in other
More informationClustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic
Clustering SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering is one of the fundamental and ubiquitous tasks in exploratory data analysis a first intuition about the
More informationLecture 4 Hierarchical clustering
CSE : Unsupervised learning Spring 00 Lecture Hierarchical clustering. Multiple levels of granularity So far we ve talked about the k-center, k-means, and k-medoid problems, all of which involve pre-specifying
More informationClustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation
More informationCluster Analysis. Ying Shen, SSE, Tongji University
Cluster Analysis Ying Shen, SSE, Tongji University Cluster analysis Cluster analysis groups data objects based only on the attributes in the data. The main objective is that The objects within a group
More informationHierarchical Clustering
Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram A tree like diagram that records the sequences of merges or splits 0 0 0 00
More informationIntroduction to spectral clustering
Introduction to spectral clustering Vasileios Zografos zografos@isy.liu.se Klas Nordberg klas@isy.liu.se What this course is Basic introduction into the core ideas of spectral clustering Sufficient to
More informationData Mining Cluster Analysis: Basic Concepts and Algorithms. Slides From Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Slides From Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining
More informationClustering Lecture 3: Hierarchical Methods
Clustering Lecture 3: Hierarchical Methods Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced
More informationLecture Notes for Chapter 7. Introduction to Data Mining, 2 nd Edition. by Tan, Steinbach, Karpatne, Kumar
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 7 Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar Hierarchical Clustering Produces a set
More informationFoundations of Computer Science Spring Mathematical Preliminaries
Foundations of Computer Science Spring 2017 Equivalence Relation, Recursive Definition, and Mathematical Induction Mathematical Preliminaries Mohammad Ashiqur Rahman Department of Computer Science College
More informationOptimization I : Brute force and Greedy strategy
Chapter 3 Optimization I : Brute force and Greedy strategy A generic definition of an optimization problem involves a set of constraints that defines a subset in some underlying space (like the Euclidean
More informationHierarchical Clustering: Objectives & Algorithms. École normale supérieure & CNRS
Hierarchical Clustering: Objectives & Algorithms Vincent Cohen-Addad Paris Sorbonne & CNRS Frederik Mallmann-Trenn MIT Varun Kanade University of Oxford Claire Mathieu École normale supérieure & CNRS Clustering
More informationGeneralized Transitive Distance with Minimum Spanning Random Forest
Generalized Transitive Distance with Minimum Spanning Random Forest Author: Zhiding Yu and B. V. K. Vijaya Kumar, et al. Dept of Electrical and Computer Engineering Carnegie Mellon University 1 Clustering
More informationCluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1
Cluster Analysis Mu-Chun Su Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Introduction Cluster analysis is the formal study of algorithms and methods
More informationClustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search
Informal goal Clustering Given set of objects and measure of similarity between them, group similar objects together What mean by similar? What is good grouping? Computation time / quality tradeoff 1 2
More informationClustering CS 550: Machine Learning
Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf
More informationData Mining Concepts & Techniques
Data Mining Concepts & Techniques Lecture No 08 Cluster Analysis Naeem Ahmed Email: naeemmahoto@gmailcom Department of Software Engineering Mehran Univeristy of Engineering and Technology Jamshoro Outline
More informationAn Axiomatic Approach to Routing. Omer Lev, Moshe Tennenholtz and Aviv Zohar TARK 2015
An Axiomatic Approach to Routing Omer Lev, Moshe Tennenholtz and Aviv Zohar TARK 205 What is routing? A network of connected nodes with edge weights 5 3 2 3 4 What is routing? A network of connected nodes
More informationCS7267 MACHINE LEARNING
S7267 MAHINE LEARNING HIERARHIAL LUSTERING Ref: hengkai Li, Department of omputer Science and Engineering, University of Texas at Arlington (Slides courtesy of Vipin Kumar) Mingon Kang, Ph.D. omputer Science,
More informationClustering. CS294 Practical Machine Learning Junming Yin 10/09/06
Clustering CS294 Practical Machine Learning Junming Yin 10/09/06 Outline Introduction Unsupervised learning What is clustering? Application Dissimilarity (similarity) of objects Clustering algorithm K-means,
More informationA Computational Theory of Clustering
A Computational Theory of Clustering Avrim Blum Carnegie Mellon University Based on work joint with Nina Balcan, Anupam Gupta, and Santosh Vempala Point of this talk A new way to theoretically analyze
More informationBBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler
BBS654 Data Mining Pinar Duygulu Slides are adapted from Nazli Ikizler 1 Classification Classification systems: Supervised learning Make a rational prediction given evidence There are several methods for
More informationCS264: Beyond Worst-Case Analysis Lecture #6: Perturbation-Stable Clustering
CS264: Beyond Worst-Case Analysis Lecture #6: Perturbation-Stable Clustering Tim Roughgarden January 26, 2017 1 Clustering Is Hard Only When It Doesn t Matter In some optimization problems, the objective
More informationChapter 6: Cluster Analysis
Chapter 6: Cluster Analysis The major goal of cluster analysis is to separate individual observations, or items, into groups, or clusters, on the basis of the values for the q variables measured on each
More informationHierarchical Graph Clustering: Quality Metrics & Algorithms
Hierarchical Graph Clustering: Quality Metrics & Algorithms Thomas Bonald Joint work with Bertrand Charpentier, Alexis Galland & Alexandre Hollocou LTCI Data Science seminar March 2019 Motivation Clustering
More informationHierarchical Clustering
What is clustering Partitioning of a data set into subsets. A cluster is a group of relatively homogeneous cases or observations Hierarchical Clustering Mikhail Dozmorov Fall 2016 2/61 What is clustering
More informationClustering Part 3. Hierarchical Clustering
Clustering Part Dr Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville Hierarchical Clustering Two main types: Agglomerative Start with the points
More informationRandomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees
Randomized Algorithms 2017A - Lecture 10 Metric Embeddings into Random Trees Lior Kamma 1 Introduction Embeddings and Distortion An embedding of a metric space (X, d X ) into a metric space (Y, d Y ) is
More informationUnsupervised Learning and Clustering
Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)
More informationUnderstanding Clustering Supervising the unsupervised
Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data
More informationCS 534: Computer Vision Segmentation and Perceptual Grouping
CS 534: Computer Vision Segmentation and Perceptual Grouping Ahmed Elgammal Dept of Computer Science CS 534 Segmentation - 1 Outlines Mid-level vision What is segmentation Perceptual Grouping Segmentation
More informationMinimum Spanning Trees
Minimum Spanning Trees Overview Problem A town has a set of houses and a set of roads. A road connects and only houses. A road connecting houses u and v has a repair cost w(u, v). Goal: Repair enough (and
More informationClustering Algorithms for general similarity measures
Types of general clustering methods Clustering Algorithms for general similarity measures general similarity measure: specified by object X object similarity matrix 1 constructive algorithms agglomerative
More informationClustering algorithms and introduction to persistent homology
Foundations of Geometric Methods in Data Analysis 2017-18 Clustering algorithms and introduction to persistent homology Frédéric Chazal INRIA Saclay - Ile-de-France frederic.chazal@inria.fr Introduction
More informationScalable Clustering of Signed Networks Using Balance Normalized Cut
Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.
More informationRepresentation Learning for Clustering: A Statistical Framework
Representation Learning for Clustering: A Statistical Framework Hassan Ashtiani School of Computer Science University of Waterloo mhzokaei@uwaterloo.ca Shai Ben-David School of Computer Science University
More informationCluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010
Cluster Analysis Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX 7575 April 008 April 010 Cluster Analysis, sometimes called data segmentation or customer segmentation,
More informationGraphs and Network Flows IE411. Lecture 21. Dr. Ted Ralphs
Graphs and Network Flows IE411 Lecture 21 Dr. Ted Ralphs IE411 Lecture 21 1 Combinatorial Optimization and Network Flows In general, most combinatorial optimization and integer programming problems are
More informationBMI/STAT 768: Lecture 06 Trees in Graphs
BMI/STAT 768: Lecture 06 Trees in Graphs Moo K. Chung mkchung@wisc.edu February 11, 2018 Parts of this lecture is based on [3, 5]. Many objects and data can be represented as networks. Unfortunately networks
More information10. Clustering. Introduction to Bioinformatics Jarkko Salojärvi. Based on lecture slides by Samuel Kaski
10. Clustering Introduction to Bioinformatics 30.9.2008 Jarkko Salojärvi Based on lecture slides by Samuel Kaski Definition of a cluster Typically either 1. A group of mutually similar samples, or 2. A
More informationLecture 5: Duality Theory
Lecture 5: Duality Theory Rajat Mittal IIT Kanpur The objective of this lecture note will be to learn duality theory of linear programming. We are planning to answer following questions. What are hyperplane
More information4. Ad-hoc I: Hierarchical clustering
4. Ad-hoc I: Hierarchical clustering Hierarchical versus Flat Flat methods generate a single partition into k clusters. The number k of clusters has to be determined by the user ahead of time. Hierarchical
More informationSTATS306B STATS306B. Clustering. Jonathan Taylor Department of Statistics Stanford University. June 3, 2010
STATS306B Jonathan Taylor Department of Statistics Stanford University June 3, 2010 Spring 2010 Outline K-means, K-medoids, EM algorithm choosing number of clusters: Gap test hierarchical clustering spectral
More information10701 Machine Learning. Clustering
171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among
More informationHierarchical Clustering
Hierarchical Clustering Hierarchical Clustering Produces a set of nested clusters organized as a hierarchical tree Can be visualized as a dendrogram A tree-like diagram that records the sequences of merges
More informationOn the Relationships between Zero Forcing Numbers and Certain Graph Coverings
On the Relationships between Zero Forcing Numbers and Certain Graph Coverings Fatemeh Alinaghipour Taklimi, Shaun Fallat 1,, Karen Meagher 2 Department of Mathematics and Statistics, University of Regina,
More informationThe Encoding Complexity of Network Coding
The Encoding Complexity of Network Coding Michael Langberg Alexander Sprintson Jehoshua Bruck California Institute of Technology Email: mikel,spalex,bruck @caltech.edu Abstract In the multicast network
More informationMatching Algorithms. Proof. If a bipartite graph has a perfect matching, then it is easy to see that the right hand side is a necessary condition.
18.433 Combinatorial Optimization Matching Algorithms September 9,14,16 Lecturer: Santosh Vempala Given a graph G = (V, E), a matching M is a set of edges with the property that no two of the edges have
More informationWhat to come. There will be a few more topics we will cover on supervised learning
Summary so far Supervised learning learn to predict Continuous target regression; Categorical target classification Linear Regression Classification Discriminative models Perceptron (linear) Logistic regression
More informationSection 3 Default Logic. Subsection 3.1 Introducing defaults and default logics
Section 3 Default Logic Subsection 3.1 Introducing defaults and default logics TU Dresden, WS 2017/18 Introduction to Nonmonotonic Reasoning Slide 34 Introducing defaults and default logics: an example
More informationCluster Analysis for Microarray Data
Cluster Analysis for Microarray Data Seventh International Long Oligonucleotide Microarray Workshop Tucson, Arizona January 7-12, 2007 Dan Nettleton IOWA STATE UNIVERSITY 1 Clustering Group objects that
More information2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006
2386 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 6, JUNE 2006 The Encoding Complexity of Network Coding Michael Langberg, Member, IEEE, Alexander Sprintson, Member, IEEE, and Jehoshua Bruck,
More informationReading 1 : Introduction
CS/Math 240: Introduction to Discrete Mathematics Fall 2015 Instructors: Beck Hasti and Gautam Prakriya Reading 1 : Introduction Welcome to CS 240, an introduction to discrete mathematics. This reading
More informationDATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm
DATA MINING LECTURE 7 Hierarchical Clustering, DBSCAN The EM Algorithm CLUSTERING What is a Clustering? In general a grouping of objects such that the objects in a group (cluster) are similar (or related)
More informationTypes of general clustering methods. Clustering Algorithms for general similarity measures. Similarity between clusters
Types of general clustering methods Clustering Algorithms for general similarity measures agglomerative versus divisive algorithms agglomerative = bottom-up build up clusters from single objects divisive
More informationClustering. Chapter 10 in Introduction to statistical learning
Clustering Chapter 10 in Introduction to statistical learning 16 14 12 10 8 6 4 2 0 2 4 6 8 10 12 14 1 Clustering ² Clustering is the art of finding groups in data (Kaufman and Rousseeuw, 1990). ² What
More informationECS 234: Data Analysis: Clustering ECS 234
: Data Analysis: Clustering What is Clustering? Given n objects, assign them to groups (clusters) based on their similarity Unsupervised Machine Learning Class Discovery Difficult, and maybe ill-posed
More informationData Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining
Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining 4/18/004 1
More informationUnsupervised Learning and Clustering
Unsupervised Learning and Clustering Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Spring 2008 CS 551, Spring 2008 c 2008, Selim Aksoy (Bilkent University)
More informationLesson 3. Prof. Enza Messina
Lesson 3 Prof. Enza Messina Clustering techniques are generally classified into these classes: PARTITIONING ALGORITHMS Directly divides data points into some prespecified number of clusters without a hierarchical
More informationA synchronizer generates sequences of clock pulses at each node of the network satisfying the condition given by the following definition.
Chapter 8 Synchronizers So far, we have mainly studied synchronous algorithms because generally, asynchronous algorithms are often more di cult to obtain and it is substantially harder to reason about
More informationINF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22
INF4820 Clustering Erik Velldal University of Oslo Nov. 17, 2009 Erik Velldal INF4820 1 / 22 Topics for Today More on unsupervised machine learning for data-driven categorization: clustering. The task
More informationCSE 5243 INTRO. TO DATA MINING
CSE 5243 INTRO. TO DATA MINING Cluster Analysis: Basic Concepts and Methods Huan Sun, CSE@The Ohio State University Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han 2 Chapter 10. Cluster
More informationRepresenting Product Designs Using a Description Graph Extension to OWL 2
Representing Product Designs Using a Description Graph Extension to OWL 2 Henson Graves Lockheed Martin Aeronautics Company Fort Worth Texas, USA henson.graves@lmco.com Abstract. Product development requires
More informationClustering. (Part 2)
Clustering (Part 2) 1 k-means clustering 2 General Observations on k-means clustering In essence, k-means clustering aims at minimizing cluster variance. It is typically used in Euclidean spaces and works
More informationThoughts on clustering
Thoughts on clustering Avrim Blum Carnegie Mellon University Abstract Clustering is a somewhat confusing topic theoretically. In large part this is because there are many different kinds of clustering
More informationNotes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017)
1 Notes Reminder: HW2 Due Today by 11:59PM TA s note: Please provide a detailed ReadMe.txt file on how to run the program on the STDLINUX. If you installed/upgraded any package on STDLINUX, you should
More informationData Clustering. Danushka Bollegala
Data Clustering Danushka Bollegala Outline Why cluster data? Clustering as unsupervised learning Clustering algorithms k-means, k-medoids agglomerative clustering Brown s clustering Spectral clustering
More informationOn Constraint Problems with Incomplete or Erroneous Data
On Constraint Problems with Incomplete or Erroneous Data Neil Yorke-Smith and Carmen Gervet IC Parc, Imperial College, London, SW7 2AZ, U.K. nys,cg6 @icparc.ic.ac.uk Abstract. Real-world constraint problems
More informationAlgorithms, Games, and Networks February 21, Lecture 12
Algorithms, Games, and Networks February, 03 Lecturer: Ariel Procaccia Lecture Scribe: Sercan Yıldız Overview In this lecture, we introduce the axiomatic approach to social choice theory. In particular,
More informationHierarchical clustering
Hierarchical clustering Based in part on slides from textbook, slides of Susan Holmes December 2, 2012 1 / 1 Description Produces a set of nested clusters organized as a hierarchical tree. Can be visualized
More informationCluster analysis. Agnieszka Nowak - Brzezinska
Cluster analysis Agnieszka Nowak - Brzezinska Outline of lecture What is cluster analysis? Clustering algorithms Measures of Cluster Validity What is Cluster Analysis? Finding groups of objects such that
More informationCluster Analysis. Angela Montanari and Laura Anderlucci
Cluster Analysis Angela Montanari and Laura Anderlucci 1 Introduction Clustering a set of n objects into k groups is usually moved by the aim of identifying internally homogenous groups according to a
More informationMSA220 - Statistical Learning for Big Data
MSA220 - Statistical Learning for Big Data Lecture 13 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Clustering Explorative analysis - finding groups
More informationUC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 8 Lecturer: David Wagner February 20, Notes 8 for CS 170
UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 8 Lecturer: David Wagner February 20, 2003 Notes 8 for CS 170 1 Minimum Spanning Trees A tree is an undirected graph that is connected
More informationCS570: Introduction to Data Mining
CS570: Introduction to Data Mining Scalable Clustering Methods: BIRCH and Others Reading: Chapter 10.3 Han, Chapter 9.5 Tan Cengiz Gunay, Ph.D. Slides courtesy of Li Xiong, Ph.D., 2011 Han, Kamber & Pei.
More informationTheorem 2.9: nearest addition algorithm
There are severe limits on our ability to compute near-optimal tours It is NP-complete to decide whether a given undirected =(,)has a Hamiltonian cycle An approximation algorithm for the TSP can be used
More informationThroughout the chapter, we will assume that the reader is familiar with the basics of phylogenetic trees.
Chapter 7 SUPERTREE ALGORITHMS FOR NESTED TAXA Philip Daniel and Charles Semple Abstract: Keywords: Most supertree algorithms combine collections of rooted phylogenetic trees with overlapping leaf sets
More informationDiscrete mathematics
Discrete mathematics Petr Kovář petr.kovar@vsb.cz VŠB Technical University of Ostrava DiM 470-2301/02, Winter term 2018/2019 About this file This file is meant to be a guideline for the lecturer. Many
More informationLecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1
CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanford.edu) January 11, 2018 Lecture 2 - Graph Theory Fundamentals - Reachability and Exploration 1 In this lecture
More informationGreedy Algorithms Part Three
Greedy Algorithms Part Three Announcements Problem Set Four due right now. Due on Wednesday with a late day. Problem Set Five out, due Monday, August 5. Explore greedy algorithms, exchange arguments, greedy
More informationInductive datatypes in HOL. lessons learned in Formal-Logic Engineering
Inductive datatypes in HOL lessons learned in Formal-Logic Engineering Stefan Berghofer and Markus Wenzel Institut für Informatik TU München = Isabelle λ β HOL α 1 Introduction Applications of inductive
More informationOn graphs with disjoint dominating and 2-dominating sets
On graphs with disjoint dominating and 2-dominating sets 1 Michael A. Henning and 2 Douglas F. Rall 1 Department of Mathematics University of Johannesburg Auckland Park, 2006 South Africa Email: mahenning@uj.ac.za
More informationTree Models of Similarity and Association. Clustering and Classification Lecture 5
Tree Models of Similarity and Association Clustering and Lecture 5 Today s Class Tree models. Hierarchical clustering methods. Fun with ultrametrics. 2 Preliminaries Today s lecture is based on the monograph
More informationCollection Guiding: Multimedia Collection Browsing and Visualization. Outline. Context. Searching for data
Collection Guiding: Multimedia Collection Browsing and Visualization Stéphane Marchand-Maillet Viper CVML University of Geneva marchand@cui.unige.ch http://viper.unige.ch Outline Multimedia data context
More informationMotivation. Technical Background
Handling Outliers through Agglomerative Clustering with Full Model Maximum Likelihood Estimation, with Application to Flow Cytometry Mark Gordon, Justin Li, Kevin Matzen, Bryce Wiedenbeck Motivation Clustering
More informationIT-Dendrogram: A New Member of the In-Tree (IT) Clustering Family
IT-Dendrogram: A New Member of the In-Tree (IT) Clustering Family Teng Qiu (qiutengcool@163.com) Yongjie Li (liyj@uestc.edu.cn) University of Electronic Science and Technology of China, Chengdu, China
More informationMa/CS 6b Class 7: Minors
Ma/CS 6b Class 7: Minors By Adam Sheffer Edge Subdivision iven a graph = V, E, and an edge e E, subdividing e is the operation of replacing e with a path consisting of new vertices. 1 raph Relations We
More informationK-means and Hierarchical Clustering
K-means and Hierarchical Clustering Xiaohui Xie University of California, Irvine K-means and Hierarchical Clustering p.1/18 Clustering Given n data points X = {x 1, x 2,, x n }. Clustering is the partitioning
More informationTrees. 3. (Minimally Connected) G is connected and deleting any of its edges gives rise to a disconnected graph.
Trees 1 Introduction Trees are very special kind of (undirected) graphs. Formally speaking, a tree is a connected graph that is acyclic. 1 This definition has some drawbacks: given a graph it is not trivial
More information