Clustering. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

Size: px
Start display at page:

Download "Clustering. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238"

Transcription

1 Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

2 What is Clustering? Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

3 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 What is Clustering? Clustering Class discovery Given a set of objects, group them into clusters (classes that are unknown beforehand) an instance of unsupervised learning (no training dataset) In Practice Cluster images to find categories Cluster patient data to find disease subtypes Cluster persons in social networks to detect communities

4 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 What is Clustering? Supervised versus unsupervised learning general inference problem: given x i, predict y i by learning a function f training set: set of examples (x i, y i ) where y i = f (x i ) (but f is still unknown!) test set: new set of data points x i where y i is unknown Supervised: use training data to infer your model, then apply this model to the test data Unsupervised: no training data, learn model and apply it directly on the test data

5 k-means Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

6 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Objective Partition the dataset into k clusters such that intra-cluster variance is minimised V (D) = k i=1 x j S i (x j µ i ) 2 (1) where V is the variance, S i is a cluster, µ i is its mean, D is the dataset of all points x j

7 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Lloyds algorithm (Lloyds, 1957) 1 Partition the data into k initial clusters 2 Compute the mean of each cluster 3 Assign each point to the cluster whose mean is closest to the point 4 If any point changed its cluster membership: Repeat from Step 2

8 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Example Example: Iteration 1

9 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Example Example: Iteration 2

10 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Example Example: Iteration 3

11 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Example Example: Iteration 4

12 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Number of Clusters Example: k = 2

13 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Number of Clusters Example: k = 3

14 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Number of Clusters Example: k = 4

15 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Number of Clusters Example: k = 5

16 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Number of Clusters Example: k = 6

17 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 1, Iteration 1

18 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 1, Iteration 5

19 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 1, Iteration 9

20 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 2, Iteration 1

21 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 2, Iteration 2

22 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 2, Iteration 3

23 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Effect of Initialization Example: Initialization 2, Iteration 4

24 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means Things to note k-means is still the state-of-the-art method for most clustering tasks When proposing a new clustering method, one should always compare to k-means. Lloyds algorithm has several setbacks It is order-dependent. Its results depends on the initialisation of the clusters. Its result may be a local optimum, not the global optimal solution.

25 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-centroid Brother of k-means Don t use the mean of each cluster but the medoid. The medoid is the point closest to the mean: m i = argmin xj S i x j µ i 2 One thereby restricts the cluster means to points that are present in the dataset. One only minimises variance with respect to these points.

26 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Kernel k-means Kernelised k-means? It would be attractive to perform clustering using kernels can move clustering problem to different feature spaces can cluster string and graph data But we have to be able to perform all steps in k-means using kernels!

27 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Kernel k-means Kernelised k-means The key step in k-means is to compute the distance between one data point x 1 and the mean of a cluster of points x 2,..., x m : φ(x 1 ) k(x 1, x 1 ) 1 (m 1) m φ(x j ) 2 = j=2 2 (m 1) m 1 k(x 1, x j ) + (m 1) 2 j=2 m i=2 j=2 This result is based on the fact that every kernel k induces a distance d: d(x i, x j ) 2 = φ(x i ) φ(x j ) 2 =k(x i, x i ) 2k(x i, x j )+k(x j, x j ) m k(x i, x j ) (2)

28 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 k-means: Silhouette Coefficients How to set k in k-means A silhouette coefficient s(x) (Rousseeuw, 1987) relates the average distance between a point x and and all others points from its cluster C, d(x, µ C ), to the average distance between a point x and the other points from the second nearest cluster C, d(x, µ C ): s(x) = d(x, µ C ) d(x, µ C ) max(d(x, µ C ), d(x, µ C )) Interpretation of s(x) : s(x) is close to 1, if a point is clearly located in its cluster C. s(x) is close to 0, if a point is located between two clusters. s(x) is negative, if it is closer to cluster C than to its current cluster.

29 k-means: Silhouette Coefficients Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

30 k-means: Silhouette Coefficients Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

31 k-means: Silhouette Coefficients Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

32 Graph-based Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

33 Graph-based Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

34 Graph-based Clustering Data representation dataset D is given in terms of a graph G = (V, E) a data object v i is a node in G edge e ij from node v i to node v j has weight w ij Graph-based clustering Define a threshold θ Remove all edges e ij from G with weight w ij > θ. Each connected component of the graph now corresponds to one cluster. Two nodes are in the same connected component if there is a path between them. Graph components can be found by depth-first search in a graph (O( V + E )). Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

35 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Graph-based Clustering Original graph

36 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Graph-based Clustering Thresholded graph (θ = 0.5)

37 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Graph-based Clustering But how to get the graph in the first place? Some domains have a natural graph structure, e.g. telecommunication or social networks Otherwise, obtain the graph structure through a distance function d on the vertices by: connecting each node to its k-nearest neighbors, connecting each node to all nodes in an ɛ-neighborhood.

38 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan Noise-robust graph-based clustering Graph-based clustering can suffer from the fact that one noisy edge connects two clusters. DBScan (Ester et al., 1996) is a noise-robust extension of graph-based clustering. DBScan is short for Density-Based Spatial Clustering of Applications with Noise. Core object Two objects v i and v j with distance d(v i, v j ) < ɛ belong to the same cluster if either v i or v j are a core object. v i is a core object iff there are MinPoints points within a distance of ɛ from v i. A cluster is defined by iteratively checking this core object property.

39 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan DBScan-clustered graph (MinPts = 2, Eps = 0.5)

40 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan DBScan-clustered graph (MinPts = 3, Eps = 0.5)

41 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan DBScan-clustered graph (MinPts = 3, Eps = 0.5)

42 DBScan Pseudocode I Code: Main DBSCAN (SetOfPoints, Eps, MinPts) // SetOfPoints is UNCLASSIFIED ClusterId := nextid(noise); for i FROM 1 TO SetOfPoints.size do Point := SetOfPoints.get(i); if Point.ClId = UNCLASSIFIED then if ExpandCluster(SetOfPoints, Point, ClusterId, Eps, MinPts) then ClusterId := nextid(clusterid) Code: ExpandCluster ExpandCluster(SetOfPoints, Point, ClId, Eps, MinPts): Boolean; seeds:=setofpoints.regionquery(point,eps); if seeds.size < MinPts then SetOfPoint.changeClId(Point,NOISE); Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

43 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan Pseudocode II RETURN False; else SetOfPoints.changeClIds(seeds,ClId); seeds.delete(point); while seeds <> Empty currentp := seeds.first(); result := SetOfPoints.regionQuery(currentP, Eps); if result.size >= MinPts then for i FROM 1 TO result.size do resultp := result.get(i); if resultp.clid IN (UNCLASSIFIED, NOISE) then if resultp.clid = UNCLASSIFIED then seeds.append(resultp); end if

44 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan Pseudocode III SetOfPoints.changeClId(resultP,ClId); end if // UNCLASSIFIED or NOISE end for; end if; // result.size >= MinPts seeds.delete(currentp); end while; // seeds <> Empty RETURN True; end if end // ExpandCluster

45 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 DBScan Properties Cluster assignment of border points is order-dependent Unlike k-means, one does not have to specify the number of clusters a priori But one has to set MinPts and Eps Ester et al. report that for 2D examples MinPts=4 is sufficient for good results They determine Eps by visual inspection of a k-distance plot Transfer question: How to kernelise DBScan?

46 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering based on: Mohammed J. Zaki and Wagner Meira Jr. Data Mining and Analysis: Fundamental Concepts and Algorithms, Cambridge University Press (2014), Chapter 16

47 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Concept Spectral Clustering connects graph-based clustering with k-means. But what is the link between these two approaches? To understand this link, we must first familiarize ourselves with cut-based clustering.

48 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Cut-based clustering 0.7$ 0.8$ 0.8$ 0.2$ 0.1$ 0.2$ 0.9$ 0.8$ 0.8$ 0.8$ 0.9$ 0.7$

49 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Cut-based clustering Objects are nodes in a graph G with nodes V and edges E. The adjacency matrix W represents the similarities between pairs of nodes. Assume V is partitioned into k subsets: V = {C 1,..., C k }. Cut-based clustering tries to minimize the total weight of inter-cluster edges: min 1 2 k a=1 b=1 k κ(c a, C b ), where κ(c a, C b ) = v i C a,v j C b,a b W ij and κ(c a, C a ) = 0.

50 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Link to the Graph Laplacian The degree matrix D is defined as D ij = { n j=1 W ij i = j 0 i j The (unnormalized) Graph Laplacian is defined as L = D W.

51 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Link to the Graph Laplacian { 1 if vi C Let c a be a vector of size n such that c a (i) = a 0 if v i / C a Note that c a, c a = c a 2 = C a, the size of cluster C a. Furthermore, note that all c a and c b are orthogonal, that is, c a, c b = 0.

52 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Link to the Graph Laplacian c a Lc a = c a (D W )c a = (3) = c a (i)c a (j)(d ij W ij ) = (4) i,j = n c a (i)( W il ) c a (i)c a (j)w ij = (5) i l=1 i,j = W ij W ij = W ij. (6) v i C a,v j C b v i C a,v j C a v i C a,v j C b,a b = κ(c a, C b ) (7) b

53 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Link to the Graph Laplacian It follows that finding the minimum k-cut is identical to minimizing min 1 2 k c a Lc a a=1

54 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Ratio Cut Minimum k cut clustering is prone to finding very small clusters. Ratio Cut accounts for this problem by dividing the cut size by the size of the cluster: min C k a=1 1 C a k κ(c a, C b ) b=1 This can be rewritten as min C k a=1 c a Lc a c a 2.

55 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Ratio Cut Finding the optimal C, that is the optimal binary cluster indicator vectors c a for a {1,..., k}, is NP-hard. We therefore relax the solution and allow the vectors c a to take any real value, rather than being binary. The objective can then be rewritten as where u a = k ( c a c a ) L( c a c a ) = a=1 k u a Lu a, (8) a=1 ca c a is the unit vector in the direction of vector c a R n.

56 Spectral Clustering Ratio Cut To find the optimal u, we first incorporate the constraint u a u a = 1 via a Lagrange multiplier λ into Objective (8): k u a Lu a + a=1 k λ a (1 u a u a ) a=1 Setting the derivative with respect to u i to zero implies: u i ( k u a Lu a + a=1 k λ a (1 u a u a )) = 0 (9) a=1 2Lu i 2λ i u i = 0, (10) Lu i = λ i u i (11) Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

57 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Ratio Cut This implies that all u a are eigenvectors of L. Using Equation (11), it follows that u a Lu a = u a λ a u a = λ a This in turn implies that in order to minimize Objective (8), one should choose the k smallest eigenvalues of L and their corresponding eigenvectors. The eigenvectors represent the relaxed cluster indicator vectors (excluding u n ).

58 Spectral Clustering Concept (Donath and Hoffman, 1973, Shi and Malik, 2000 and Ng, Jordan, and Weiss, 2002) Spectral Clustering solves this problem pragmatically by using the vectors u a as a new representation of the data points and applying k-means to this new representation after normalization. The new representation is U = (u n, u n 1... u n k+1 ). It is normalized row-by-row to obtain the new k-dimensional representation y 1 Y = y2..., via y 1 i = k (u n,i, u n 1,i,..., u n k+1,i ). (12) yn j=1 u2 n j+1,i Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

59 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Pseudocode procedure Spectral Clustering(D, k) Compute the similarity matrix W R n n and the Laplacian L = D W ; Solve Lu a = λ a u a for a = n,..., n k + 1, where λ n λ n 1... λ n k+1 U := (u n, u n 1,..., u n k+1 ) Y := normalized rows of U via Equation (12) C := {C 1,..., C k } via k-means on Y return: C

60 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Spectral Clustering Computational Complexity The overall worst case runtime is O(n 3 ) due to the need to computer eigenvectors and eigenvalues. For sparse graphs with m edges, this runtime can be improved to O(mn). Running k-means requires a runtime in O(tnk 2 ), where t is the number of iterations of k-means until convergence.

61 Soft-assignment Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

62 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering Soft k-means k-means is based on a hard assignment of points to clusters. Wouldn t it be more realistic to work with the probabilities of each point to belong to each cluster rather than a hard assignment? This is the core idea of Expectation Maximization (EM) Clustering with a Mixture of Gaussian distributions. It is also referred to as soft k-means.

63 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering The General EM algorithm (Dempster et al., 1977) We are dealing with observed variables (objects X and their features) and latent variables (the cluster membership of the objects Y ), and model parameters θ (parameters of the underlying probability distribution). We would like to maximize p(x θ). This optimization is difficult as log p(x θ) = log{ Y p(x, Y θ)}, that is, we have to sum over the latent variables inside the logarithm, which makes the evaluation of the maximum likelihood extremely challenging. The EM algorithm circumvents this problem in an iterative 2-step procedure.

64 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering The General EM algorithm (Dempster et al., 1977; Source: Bishop, 2006, Chapter 9.2) Given a joint distribution p(x, Y θ) over observed variables X and latent variables Y, governed by parameters θ, the goal is to maximize the likelihood function p(x θ) with respect to θ: 1 Choose an initial setting for the parameters θ old. 2 Expectation step (E step): Evaluate p(y X, θ old ). 3 Maximization step (M step): Evaluate θ new given by θ new = arg max Q(θ, θ old ), θ where Q(θ, θ old ) = Y p(y X, θold ) log p(x, Y θ). 4 Check for convergence of parameters or log likelihood. If not converged, then θ old θ new and return to Step 2. EM may converge to a local optimum.

65 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering Concept Consider the case of a mixture of k Gaussians in which θ is a triplet (c, {µ 1,..., µ k }, {Σ 1,..., Σ k }), where P θ [Y = y] = c y and P θ (t)[x = x Y = y] = N (x µ y, Σ y ) = 1 (2π) d 2 Σ y 1 2 e 1 2 ((x µ y ) Σ 1 y (x µ y )) For simplicity, we assume that Σ 1 = Σ 2 =... = Σ k = I, where I is the identity matrix.

66 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering Expectation step For each i {1,..., n} and y {1,..., k}, we have that P θ (t)[y = y X = x i ] = 1 P Z θ (t)[y = y]p θ (t)[x = x i Y = y] = 1 c y (t) exp( 1 i Z i 2 x i µ (t) y 2 ), where Z i is a normalization factor which ensures that Σ y P θ (t)[y = y X = x i ] sums to 1.

67 EM Clustering Maximization step We need to maximize the following expression with respect to µ and c to obtain θ (t+1) : n k P θ (t)[y = y X = x i ](log(c y ) 1 2 x i µ y 2 ). i=1 y=1 Setting the derivative wrt µ y of the term above to zero and rearranging terms, one obtains: µ y = n i=1 P θ (t)[y = y X = x i]x i n i=1 P θ (t)[y = y X = x i] µ y is the weighted average of the x i, where the weights are according to the probabilities calculated in the E step. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

68 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering Maximization step The maximal c can be shown to be c y = n i=1 P θ (t)[y = y X = x i] k n y =1 i=1 P θ (t)[y = y X = x i ]

69 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 EM Clustering Comparison to k-means k-means assigns each point to a cluster according to the distance to the cluster means. Then the cluster means are updated based on the examples in this cluster. EM determines the probability that each example belongs to each cluster. Then the cluster means are updated based on a weighted sum over all data points.

70 Hierarchical Clustering Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238

71 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Hierarchical Clustering Extension of original setting What if clusters contain clusters themselves? Then we need hierarchical clustering!

72 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Hierarchical Clustering Join the most similar clusters Iteratively join the two most similar clusters But how to measure similarity between clusters? Similarity of clusters (Florek et al., 1951) Single Link: s(c i, C j ) = min d(x, x ) x C i,x C j Average Link: s(c i, C j ) = 1 C i C j d(x, x ) x C i,x C j Complete Link: s(c i, C j ) = max d(x, x ) x C i,x C j

73 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Hierarchical Clustering Pseudocode procedure Hierarchical Clustering(D, s) Initialize each point x i D as its own cluster C i, for i {1,..., n}. repeat (i, j ) = arg min (i,j) s(c i, C j ) Merge clusters C i and C j. until C = 1

74 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Hierarchical Clustering Example: Dendrograms as a Way to Represent Hierarchical Clusterings

75 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Hierarchical Clustering Advantage Its clustering reflects the entire structure of the dataset. Disadvantage It is difficult to make a clear statement about cluster membership in hierarchical clustering, as each point belong to a hierarchy of clusters. Stopping Hierarchical Clustering early is an approach to circumvent this cluster-assignment problem. These criteria could be to stop the merging of clusters once a pre-defined number of clusters k has been reached. once the distance between the closest clusters exceeds a threshold ɛ.

76 Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 238 Clustering Summary Clustering partitions a dataset into groups of similar objects. The three most popular families of clustering algorithms are 1 k-means clustering, 2 graph-based clustering, 3 hierarchical clustering. When applying these algorithms, it is essential to be aware of the strengths and weaknesses of these algorithms and to report the exact parameter settings used (e.g. number of clusters, distance function used).

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering An unsupervised machine learning problem Grouping a set of objects in such a way that objects in the same group (a cluster) are more similar (in some sense or another) to each other than to those in other

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Introduction Often, we have only a set of features x = x 1, x 2,, x n, but no associated response y. Therefore we are not interested in prediction nor classification,

More information

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic

Clustering. SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic Clustering is one of the fundamental and ubiquitous tasks in exploratory data analysis a first intuition about the

More information

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm

DATA MINING LECTURE 7. Hierarchical Clustering, DBSCAN The EM Algorithm DATA MINING LECTURE 7 Hierarchical Clustering, DBSCAN The EM Algorithm CLUSTERING What is a Clustering? In general a grouping of objects such that the objects in a group (cluster) are similar (or related)

More information

Unsupervised Learning

Unsupervised Learning Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised

More information

Clustering CS 550: Machine Learning

Clustering CS 550: Machine Learning Clustering CS 550: Machine Learning This slide set mainly uses the slides given in the following links: http://www-users.cs.umn.edu/~kumar/dmbook/ch8.pdf http://www-users.cs.umn.edu/~kumar/dmbook/dmslides/chap8_basic_cluster_analysis.pdf

More information

Clustering: Classic Methods and Modern Views

Clustering: Classic Methods and Modern Views Clustering: Classic Methods and Modern Views Marina Meilă University of Washington mmp@stat.washington.edu June 22, 2015 Lorentz Center Workshop on Clusters, Games and Axioms Outline Paradigms for clustering

More information

Visual Representations for Machine Learning

Visual Representations for Machine Learning Visual Representations for Machine Learning Spectral Clustering and Channel Representations Lecture 1 Spectral Clustering: introduction and confusion Michael Felsberg Klas Nordberg The Spectral Clustering

More information

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic

Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic SEMANTIC COMPUTING Lecture 6: Unsupervised Machine Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 23 November 2018 Overview Unsupervised Machine Learning overview Association

More information

Data Mining. Clustering. Hamid Beigy. Sharif University of Technology. Fall 1394

Data Mining. Clustering. Hamid Beigy. Sharif University of Technology. Fall 1394 Data Mining Clustering Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1394 1 / 31 Table of contents 1 Introduction 2 Data matrix and

More information

Unsupervised Learning : Clustering

Unsupervised Learning : Clustering Unsupervised Learning : Clustering Things to be Addressed Traditional Learning Models. Cluster Analysis K-means Clustering Algorithm Drawbacks of traditional clustering algorithms. Clustering as a complex

More information

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A

MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A MultiDimensional Signal Processing Master Degree in Ingegneria delle Telecomunicazioni A.A. 205-206 Pietro Guccione, PhD DEI - DIPARTIMENTO DI INGEGNERIA ELETTRICA E DELL INFORMAZIONE POLITECNICO DI BARI

More information

Data Mining Algorithms

Data Mining Algorithms for the original version: -JörgSander and Martin Ester - Jiawei Han and Micheline Kamber Data Management and Exploration Prof. Dr. Thomas Seidl Data Mining Algorithms Lecture Course with Tutorials Wintersemester

More information

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme, Nicolas Schilling Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme, Nicolas Schilling Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim,

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-25-2018 Outline Background Defining proximity Clustering methods Determining number of clusters Other approaches Cluster analysis as unsupervised Learning Unsupervised

More information

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06 Clustering CS294 Practical Machine Learning Junming Yin 10/09/06 Outline Introduction Unsupervised learning What is clustering? Application Dissimilarity (similarity) of objects Clustering algorithm K-means,

More information

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme Machine Learning B. Unsupervised Learning B.1 Cluster Analysis Lars Schmidt-Thieme Information Systems and Machine Learning Lab (ISMLL) Institute for Computer Science University of Hildesheim, Germany

More information

Knowledge Discovery in Databases

Knowledge Discovery in Databases Ludwig-Maximilians-Universität München Institut für Informatik Lehr- und Forschungseinheit für Datenbanksysteme Lecture notes Knowledge Discovery in Databases Summer Semester 2012 Lecture 8: Clustering

More information

Clustering. So far in the course. Clustering. Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. dist(x, y) = x y 2 2

Clustering. So far in the course. Clustering. Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. dist(x, y) = x y 2 2 So far in the course Clustering Subhransu Maji : Machine Learning 2 April 2015 7 April 2015 Supervised learning: learning with a teacher You had training data which was (feature, label) pairs and the goal

More information

http://www.xkcd.com/233/ Text Clustering David Kauchak cs160 Fall 2009 adapted from: http://www.stanford.edu/class/cs276/handouts/lecture17-clustering.ppt Administrative 2 nd status reports Paper review

More information

Introduction to spectral clustering

Introduction to spectral clustering Introduction to spectral clustering Vasileios Zografos zografos@isy.liu.se Klas Nordberg klas@isy.liu.se What this course is Basic introduction into the core ideas of spectral clustering Sufficient to

More information

Clustering. Chapter 10 in Introduction to statistical learning

Clustering. Chapter 10 in Introduction to statistical learning Clustering Chapter 10 in Introduction to statistical learning 16 14 12 10 8 6 4 2 0 2 4 6 8 10 12 14 1 Clustering ² Clustering is the art of finding groups in data (Kaufman and Rousseeuw, 1990). ² What

More information

MSA220 - Statistical Learning for Big Data

MSA220 - Statistical Learning for Big Data MSA220 - Statistical Learning for Big Data Lecture 13 Rebecka Jörnsten Mathematical Sciences University of Gothenburg and Chalmers University of Technology Clustering Explorative analysis - finding groups

More information

Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. 2 April April 2015

Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. 2 April April 2015 Clustering Subhransu Maji CMPSCI 689: Machine Learning 2 April 2015 7 April 2015 So far in the course Supervised learning: learning with a teacher You had training data which was (feature, label) pairs

More information

Feature Selection. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 262

Feature Selection. Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester / 262 Feature Selection Department Biosysteme Karsten Borgwardt Data Mining Course Basel Fall Semester 2016 239 / 262 What is Feature Selection? Department Biosysteme Karsten Borgwardt Data Mining Course Basel

More information

Introduction to Mobile Robotics

Introduction to Mobile Robotics Introduction to Mobile Robotics Clustering Wolfram Burgard Cyrill Stachniss Giorgio Grisetti Maren Bennewitz Christian Plagemann Clustering (1) Common technique for statistical data analysis (machine learning,

More information

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University Descriptive model A descriptive model presents the main features of the data

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning Associate Instructor: Herke van Hoof (herke.vanhoof@mail.mcgill.ca) Slides mostly by: (jpineau@cs.mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551

More information

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani Clustering CE-717: Machine Learning Sharif University of Technology Spring 2016 Soleymani Outline Clustering Definition Clustering main approaches Partitional (flat) Hierarchical Clustering validation

More information

Clustering Lecture 5: Mixture Model

Clustering Lecture 5: Mixture Model Clustering Lecture 5: Mixture Model Jing Gao SUNY Buffalo 1 Outline Basics Motivation, definition, evaluation Methods Partitional Hierarchical Density-based Mixture model Spectral methods Advanced topics

More information

Segmentation: Clustering, Graph Cut and EM

Segmentation: Clustering, Graph Cut and EM Segmentation: Clustering, Graph Cut and EM Ying Wu Electrical Engineering and Computer Science Northwestern University, Evanston, IL 60208 yingwu@northwestern.edu http://www.eecs.northwestern.edu/~yingwu

More information

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013

Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Your Name: Your student id: Solution Sketches Midterm Exam COSC 6342 Machine Learning March 20, 2013 Problem 1 [5+?]: Hypothesis Classes Problem 2 [8]: Losses and Risks Problem 3 [11]: Model Generation

More information

CS Introduction to Data Mining Instructor: Abdullah Mueen

CS Introduction to Data Mining Instructor: Abdullah Mueen CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 8: ADVANCED CLUSTERING (FUZZY AND CO -CLUSTERING) Review: Basic Cluster Analysis Methods (Chap. 10) Cluster Analysis: Basic Concepts

More information

SGN (4 cr) Chapter 11

SGN (4 cr) Chapter 11 SGN-41006 (4 cr) Chapter 11 Clustering Jussi Tohka & Jari Niemi Department of Signal Processing Tampere University of Technology February 25, 2014 J. Tohka & J. Niemi (TUT-SGN) SGN-41006 (4 cr) Chapter

More information

10701 Machine Learning. Clustering

10701 Machine Learning. Clustering 171 Machine Learning Clustering What is Clustering? Organizing data into clusters such that there is high intra-cluster similarity low inter-cluster similarity Informally, finding natural groupings among

More information

Machine learning - HT Clustering

Machine learning - HT Clustering Machine learning - HT 2016 10. Clustering Varun Kanade University of Oxford March 4, 2016 Announcements Practical Next Week - No submission Final Exam: Pick up on Monday Material covered next week is not

More information

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining Data Mining Cluster Analysis: Basic Concepts and Algorithms Lecture Notes for Chapter 8 Introduction to Data Mining by Tan, Steinbach, Kumar Tan,Steinbach, Kumar Introduction to Data Mining 4/18/004 1

More information

Working with Unlabeled Data Clustering Analysis. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Working with Unlabeled Data Clustering Analysis. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan Working with Unlabeled Data Clustering Analysis Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan chanhl@mail.cgu.edu.tw Unsupervised learning Finding centers of similarity using

More information

University of Florida CISE department Gator Engineering. Clustering Part 5

University of Florida CISE department Gator Engineering. Clustering Part 5 Clustering Part 5 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville SNN Approach to Clustering Ordinary distance measures have problems Euclidean

More information

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014

Spectral Clustering. Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 Spectral Clustering Presented by Eldad Rubinstein Based on a Tutorial by Ulrike von Luxburg TAU Big Data Processing Seminar December 14, 2014 What are we going to talk about? Introduction Clustering and

More information

Machine Learning for Data Science (CS4786) Lecture 11

Machine Learning for Data Science (CS4786) Lecture 11 Machine Learning for Data Science (CS4786) Lecture 11 Spectral Clustering Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016fa/ Survey Survey Survey Competition I Out! Preliminary report of

More information

Foundations of Machine Learning CentraleSupélec Fall Clustering Chloé-Agathe Azencot

Foundations of Machine Learning CentraleSupélec Fall Clustering Chloé-Agathe Azencot Foundations of Machine Learning CentraleSupélec Fall 2017 12. Clustering Chloé-Agathe Azencot Centre for Computational Biology, Mines ParisTech chloe-agathe.azencott@mines-paristech.fr Learning objectives

More information

CSE 5243 INTRO. TO DATA MINING

CSE 5243 INTRO. TO DATA MINING CSE 5243 INTRO. TO DATA MINING Cluster Analysis: Basic Concepts and Methods Huan Sun, CSE@The Ohio State University 09/25/2017 Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han 2 Chapter 10.

More information

Clustering algorithms

Clustering algorithms Clustering algorithms Machine Learning Hamid Beigy Sharif University of Technology Fall 1393 Hamid Beigy (Sharif University of Technology) Clustering algorithms Fall 1393 1 / 22 Table of contents 1 Supervised

More information

Introduction to Machine Learning CMU-10701

Introduction to Machine Learning CMU-10701 Introduction to Machine Learning CMU-10701 Clustering and EM Barnabás Póczos & Aarti Singh Contents Clustering K-means Mixture of Gaussians Expectation Maximization Variational Methods 2 Clustering 3 K-

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Clustering Varun Chandola Computer Science & Engineering State University of New York at Buffalo Buffalo, NY, USA chandola@buffalo.edu Chandola@UB CSE 474/574 1 / 19 Outline

More information

Clustering part II 1

Clustering part II 1 Clustering part II 1 Clustering What is Cluster Analysis? Types of Data in Cluster Analysis A Categorization of Major Clustering Methods Partitioning Methods Hierarchical Methods 2 Partitioning Algorithms:

More information

Unsupervised: no target value to predict

Unsupervised: no target value to predict Clustering Unsupervised: no target value to predict Differences between models/algorithms: Exclusive vs. overlapping Deterministic vs. probabilistic Hierarchical vs. flat Incremental vs. batch learning

More information

Clustering: K-means and Kernel K-means

Clustering: K-means and Kernel K-means Clustering: K-means and Kernel K-means Piyush Rai Machine Learning (CS771A) Aug 31, 2016 Machine Learning (CS771A) Clustering: K-means and Kernel K-means 1 Clustering Usually an unsupervised learning problem

More information

Data Clustering. Danushka Bollegala

Data Clustering. Danushka Bollegala Data Clustering Danushka Bollegala Outline Why cluster data? Clustering as unsupervised learning Clustering algorithms k-means, k-medoids agglomerative clustering Brown s clustering Spectral clustering

More information

Machine Learning for OR & FE

Machine Learning for OR & FE Machine Learning for OR & FE Unsupervised Learning: Clustering Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com (Some material

More information

Clustering Part 4 DBSCAN

Clustering Part 4 DBSCAN Clustering Part 4 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville DBSCAN DBSCAN is a density based clustering algorithm Density = number of

More information

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components

( ) =cov X Y = W PRINCIPAL COMPONENT ANALYSIS. Eigenvectors of the covariance matrix are the principal components Review Lecture 14 ! PRINCIPAL COMPONENT ANALYSIS Eigenvectors of the covariance matrix are the principal components 1. =cov X Top K principal components are the eigenvectors with K largest eigenvalues

More information

Scalable Clustering of Signed Networks Using Balance Normalized Cut

Scalable Clustering of Signed Networks Using Balance Normalized Cut Scalable Clustering of Signed Networks Using Balance Normalized Cut Kai-Yang Chiang,, Inderjit S. Dhillon The 21st ACM International Conference on Information and Knowledge Management (CIKM 2012) Oct.

More information

Unsupervised Learning. Andrea G. B. Tettamanzi I3S Laboratory SPARKS Team

Unsupervised Learning. Andrea G. B. Tettamanzi I3S Laboratory SPARKS Team Unsupervised Learning Andrea G. B. Tettamanzi I3S Laboratory SPARKS Team Table of Contents 1)Clustering: Introduction and Basic Concepts 2)An Overview of Popular Clustering Methods 3)Other Unsupervised

More information

DS504/CS586: Big Data Analytics Big Data Clustering II

DS504/CS586: Big Data Analytics Big Data Clustering II Welcome to DS504/CS586: Big Data Analytics Big Data Clustering II Prof. Yanhua Li Time: 6pm 8:50pm Thu Location: AK 232 Fall 2016 More Discussions, Limitations v Center based clustering K-means BFR algorithm

More information

Expectation Maximization!

Expectation Maximization! Expectation Maximization! adapted from: Doug Downey and Bryan Pardo, Northwestern University and http://www.stanford.edu/class/cs276/handouts/lecture17-clustering.ppt Steps in Clustering Select Features

More information

Application of Spectral Clustering Algorithm

Application of Spectral Clustering Algorithm 1/27 Application of Spectral Clustering Algorithm Danielle Middlebrooks dmiddle1@math.umd.edu Advisor: Kasso Okoudjou kasso@umd.edu Department of Mathematics University of Maryland- College Park Advance

More information

Machine Learning. Unsupervised Learning. Manfred Huber

Machine Learning. Unsupervised Learning. Manfred Huber Machine Learning Unsupervised Learning Manfred Huber 2015 1 Unsupervised Learning In supervised learning the training data provides desired target output for learning In unsupervised learning the training

More information

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning Unsupervised Learning Clustering and the EM Algorithm Susanna Ricco Supervised Learning Given data in the form < x, y >, y is the target to learn. Good news: Easy to tell if our algorithm is giving the

More information

ECG782: Multidimensional Digital Signal Processing

ECG782: Multidimensional Digital Signal Processing ECG782: Multidimensional Digital Signal Processing Object Recognition http://www.ee.unlv.edu/~b1morris/ecg782/ 2 Outline Knowledge Representation Statistical Pattern Recognition Neural Networks Boosting

More information

Clustering web search results

Clustering web search results Clustering K-means Machine Learning CSE546 Emily Fox University of Washington November 4, 2013 1 Clustering images Set of Images [Goldberger et al.] 2 1 Clustering web search results 3 Some Data 4 2 K-means

More information

Clustering in Ratemaking: Applications in Territories Clustering

Clustering in Ratemaking: Applications in Territories Clustering Clustering in Ratemaking: Applications in Territories Clustering Ji Yao, PhD FIA ASTIN 13th-16th July 2008 INTRODUCTION Structure of talk Quickly introduce clustering and its application in insurance ratemaking

More information

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points]

CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, :59pm, PDF to Canvas [100 points] CIS 520, Machine Learning, Fall 2015: Assignment 7 Due: Mon, Nov 16, 2015. 11:59pm, PDF to Canvas [100 points] Instructions. Please write up your responses to the following problems clearly and concisely.

More information

Exploratory data analysis for microarrays

Exploratory data analysis for microarrays Exploratory data analysis for microarrays Jörg Rahnenführer Computational Biology and Applied Algorithmics Max Planck Institute for Informatics D-66123 Saarbrücken Germany NGFN - Courses in Practical DNA

More information

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask

Machine Learning and Data Mining. Clustering (1): Basics. Kalev Kask Machine Learning and Data Mining Clustering (1): Basics Kalev Kask Unsupervised learning Supervised learning Predict target value ( y ) given features ( x ) Unsupervised learning Understand patterns of

More information

Note Set 4: Finite Mixture Models and the EM Algorithm

Note Set 4: Finite Mixture Models and the EM Algorithm Note Set 4: Finite Mixture Models and the EM Algorithm Padhraic Smyth, Department of Computer Science University of California, Irvine Finite Mixture Models A finite mixture model with K components, for

More information

PAM algorithm. Types of Data in Cluster Analysis. A Categorization of Major Clustering Methods. Partitioning i Methods. Hierarchical Methods

PAM algorithm. Types of Data in Cluster Analysis. A Categorization of Major Clustering Methods. Partitioning i Methods. Hierarchical Methods Whatis Cluster Analysis? Clustering Types of Data in Cluster Analysis Clustering part II A Categorization of Major Clustering Methods Partitioning i Methods Hierarchical Methods Partitioning i i Algorithms:

More information

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1 Week 8 Based in part on slides from textbook, slides of Susan Holmes December 2, 2012 1 / 1 Part I Clustering 2 / 1 Clustering Clustering Goal: Finding groups of objects such that the objects in a group

More information

CHAPTER 4: CLUSTER ANALYSIS

CHAPTER 4: CLUSTER ANALYSIS CHAPTER 4: CLUSTER ANALYSIS WHAT IS CLUSTER ANALYSIS? A cluster is a collection of data-objects similar to one another within the same group & dissimilar to the objects in other groups. Cluster analysis

More information

CSE 5243 INTRO. TO DATA MINING

CSE 5243 INTRO. TO DATA MINING CSE 5243 INTRO. TO DATA MINING Cluster Analysis: Basic Concepts and Methods Huan Sun, CSE@The Ohio State University 09/28/2017 Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han 2 Chapter 10.

More information

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/09/2018)

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/09/2018) 1 Notes Reminder: HW2 Due Today by 11:59PM TA s note: Please provide a detailed ReadMe.txt file on how to run the program on the STDLINUX. If you installed/upgraded any package on STDLINUX, you should

More information

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017)

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017) 1 Notes Reminder: HW2 Due Today by 11:59PM TA s note: Please provide a detailed ReadMe.txt file on how to run the program on the STDLINUX. If you installed/upgraded any package on STDLINUX, you should

More information

Clustering in Data Mining

Clustering in Data Mining Clustering in Data Mining Classification Vs Clustering When the distribution is based on a single parameter and that parameter is known for each object, it is called classification. E.g. Children, young,

More information

K-means and Hierarchical Clustering

K-means and Hierarchical Clustering K-means and Hierarchical Clustering Xiaohui Xie University of California, Irvine K-means and Hierarchical Clustering p.1/18 Clustering Given n data points X = {x 1, x 2,, x n }. Clustering is the partitioning

More information

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a Week 9 Based in part on slides from textbook, slides of Susan Holmes Part I December 2, 2012 Hierarchical Clustering 1 / 1 Produces a set of nested clusters organized as a Hierarchical hierarchical clustering

More information

数据挖掘 Introduction to Data Mining

数据挖掘 Introduction to Data Mining 数据挖掘 Introduction to Data Mining Philippe Fournier-Viger Full professor School of Natural Sciences and Humanities philfv8@yahoo.com Spring 2019 S8700113C 1 Introduction Last week: Association Analysis

More information

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22 INF4820 Clustering Erik Velldal University of Oslo Nov. 17, 2009 Erik Velldal INF4820 1 / 22 Topics for Today More on unsupervised machine learning for data-driven categorization: clustering. The task

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Unsupervised learning Until now, we have assumed our training samples are labeled by their category membership. Methods that use labeled samples are said to be supervised. However,

More information

Mining Social Network Graphs

Mining Social Network Graphs Mining Social Network Graphs Analysis of Large Graphs: Community Detection Rafael Ferreira da Silva rafsilva@isi.edu http://rafaelsilva.com Note to other teachers and users of these slides: We would be

More information

The K-modes and Laplacian K-modes algorithms for clustering

The K-modes and Laplacian K-modes algorithms for clustering The K-modes and Laplacian K-modes algorithms for clustering Miguel Á. Carreira-Perpiñán Electrical Engineering and Computer Science University of California, Merced http://faculty.ucmerced.edu/mcarreira-perpinan

More information

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler BBS654 Data Mining Pinar Duygulu Slides are adapted from Nazli Ikizler 1 Classification Classification systems: Supervised learning Make a rational prediction given evidence There are several methods for

More information

Cluster Analysis (b) Lijun Zhang

Cluster Analysis (b) Lijun Zhang Cluster Analysis (b) Lijun Zhang zlj@nju.edu.cn http://cs.nju.edu.cn/zlj Outline Grid-Based and Density-Based Algorithms Graph-Based Algorithms Non-negative Matrix Factorization Cluster Validation Summary

More information

k-means demo Administrative Machine learning: Unsupervised learning" Assignment 5 out

k-means demo Administrative Machine learning: Unsupervised learning Assignment 5 out Machine learning: Unsupervised learning" David Kauchak cs Spring 0 adapted from: http://www.stanford.edu/class/cs76/handouts/lecture7-clustering.ppt http://www.youtube.com/watch?v=or_-y-eilqo Administrative

More information

University of Florida CISE department Gator Engineering. Clustering Part 4

University of Florida CISE department Gator Engineering. Clustering Part 4 Clustering Part 4 Dr. Sanjay Ranka Professor Computer and Information Science and Engineering University of Florida, Gainesville DBSCAN DBSCAN is a density based clustering algorithm Density = number of

More information

ECLT 5810 Clustering

ECLT 5810 Clustering ECLT 5810 Clustering What is Cluster Analysis? Cluster: a collection of data objects Similar to one another within the same cluster Dissimilar to the objects in other clusters Cluster analysis Grouping

More information

Lecture 11: E-M and MeanShift. CAP 5415 Fall 2007

Lecture 11: E-M and MeanShift. CAP 5415 Fall 2007 Lecture 11: E-M and MeanShift CAP 5415 Fall 2007 Review on Segmentation by Clustering Each Pixel Data Vector Example (From Comanciu and Meer) Review of k-means Let's find three clusters in this data These

More information

Behavioral Data Mining. Lecture 18 Clustering

Behavioral Data Mining. Lecture 18 Clustering Behavioral Data Mining Lecture 18 Clustering Outline Why? Cluster quality K-means Spectral clustering Generative Models Rationale Given a set {X i } for i = 1,,n, a clustering is a partition of the X i

More information

MATH 567: Mathematical Techniques in Data

MATH 567: Mathematical Techniques in Data Supervised and unsupervised learning Supervised learning problems: MATH 567: Mathematical Techniques in Data (X, Y ) P (X, Y ). Data Science Clustering I is labelled (input/output) with joint density We

More information

CSE 5243 INTRO. TO DATA MINING

CSE 5243 INTRO. TO DATA MINING CSE 5243 INTRO. TO DATA MINING Cluster Analysis: Basic Concepts and Methods Huan Sun, CSE@The Ohio State University Slides adapted from UIUC CS412, Fall 2017, by Prof. Jiawei Han 2 Chapter 10. Cluster

More information

DS504/CS586: Big Data Analytics Big Data Clustering II

DS504/CS586: Big Data Analytics Big Data Clustering II Welcome to DS504/CS586: Big Data Analytics Big Data Clustering II Prof. Yanhua Li Time: 6pm 8:50pm Thu Location: KH 116 Fall 2017 Updates: v Progress Presentation: Week 15: 11/30 v Next Week Office hours

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Overview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010

Overview Citation. ML Introduction. Overview Schedule. ML Intro Dataset. Introduction to Semi-Supervised Learning Review 10/4/2010 INFORMATICS SEMINAR SEPT. 27 & OCT. 4, 2010 Introduction to Semi-Supervised Learning Review 2 Overview Citation X. Zhu and A.B. Goldberg, Introduction to Semi- Supervised Learning, Morgan & Claypool Publishers,

More information

Unsupervised Learning

Unsupervised Learning Unsupervised Learning Learning without Class Labels (or correct outputs) Density Estimation Learn P(X) given training data for X Clustering Partition data into clusters Dimensionality Reduction Discover

More information

ECLT 5810 Clustering

ECLT 5810 Clustering ECLT 5810 Clustering What is Cluster Analysis? Cluster: a collection of data objects Similar to one another within the same cluster Dissimilar to the objects in other clusters Cluster analysis Grouping

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 01-31-017 Outline Background Defining proximity Clustering methods Determining number of clusters Comparing two solutions Cluster analysis as unsupervised Learning

More information

COMP 465: Data Mining Still More on Clustering

COMP 465: Data Mining Still More on Clustering 3/4/015 Exercise COMP 465: Data Mining Still More on Clustering Slides Adapted From : Jiawei Han, Micheline Kamber & Jian Pei Data Mining: Concepts and Techniques, 3 rd ed. Describe each of the following

More information

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information