MI example for poultry/export in Reuters. Overview. Introduction to Information Retrieval. Outline.

Similar documents
Introduction to Information Retrieval

Flat Clustering. Slides are mostly from Hinrich Schütze. March 27, 2017

CSE 7/5337: Information Retrieval and Web Search Document clustering I (IIR 16)

Introduction to Information Retrieval

PV211: Introduction to Information Retrieval

Introduction to Information Retrieval

Information Retrieval and Organisation

Cluster Evaluation and Expectation Maximization! adapted from: Doug Downey and Bryan Pardo, Northwestern University

Clustering CE-324: Modern Information Retrieval Sharif University of Technology

Expectation Maximization!

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering


Administrative. Machine learning code. Supervised learning (e.g. classification) Machine learning: Unsupervised learning" BANANAS APPLES

Information Retrieval and Web Search Engines

Information Retrieval and Web Search Engines

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

CS490W. Text Clustering. Luo Si. Department of Computer Science Purdue University

Chapter 9. Classification and Clustering

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

CLUSTERING. Quiz information. today 11/14/13% ! The second midterm quiz is on Thursday (11/21) ! In-class (75 minutes!)

Clustering Results. Result List Example. Clustering Results. Information Retrieval

Clustering: Overview and K-means algorithm

Text Documents clustering using K Means Algorithm

Clustering: Overview and K-means algorithm

DD2475 Information Retrieval Lecture 10: Clustering. Document Clustering. Recap: Classification. Today

Based on Raymond J. Mooney s slides

CS47300: Web Information Search and Management

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search

Data Clustering. Danushka Bollegala

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

Text classification II CE-324: Modern Information Retrieval Sharif University of Technology

Search Engines. Information Retrieval in Practice

Data Informatics. Seon Ho Kim, Ph.D.

What to come. There will be a few more topics we will cover on supervised learning

Clustering. Partition unlabeled examples into disjoint subsets of clusters, such that:

Informa(on Retrieval

PV211: Introduction to Information Retrieval

INF 4300 Classification III Anne Solberg The agenda today:

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Clust Clus e t ring 2 Nov

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Clustering (COSC 416) Nazli Goharian. Document Clustering.

Informa(on Retrieval

VECTOR SPACE CLASSIFICATION

Natural Language Processing

Network Traffic Measurements and Analysis

Clustering (COSC 488) Nazli Goharian. Document Clustering.

Big Data Analytics! Special Topics for Computer Science CSE CSE Feb 9

CS145: INTRODUCTION TO DATA MINING

Lecture on Modeling Tools for Clustering & Regression

CSE 158. Web Mining and Recommender Systems. Midterm recap

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Gene Clustering & Classification

Today s topic CS347. Results list clustering example. Why cluster documents. Clustering documents. Lecture 8 May 7, 2001 Prabhakar Raghavan

Unsupervised Learning Partitioning Methods

CSE 494/598 Lecture-11: Clustering & Classification

Clustering. Bruno Martins. 1 st Semester 2012/2013

Exploratory Analysis: Clustering

Hierarchical Clustering 4/5/17

Unsupervised Learning and Clustering

Lecture 8 May 7, Prabhakar Raghavan

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

Clustering CS 550: Machine Learning

DATA MINING INTRODUCTION TO CLASSIFICATION USING LINEAR CLASSIFIERS

Unsupervised Learning

Chapter 4: Text Clustering

Classification and Clustering

Kernels and Clustering

A novel supervised learning algorithm and its use for Spam Detection in Social Bookmarking Systems

Document Clustering: Comparison of Similarity Measures

Clustering. Supervised vs. Unsupervised Learning

Administrative. Machine learning code. Machine learning: Unsupervised learning

CS 2750 Machine Learning. Lecture 19. Clustering. CS 2750 Machine Learning. Clustering. Groups together similar instances in the data sample

Clustering Algorithms for general similarity measures

Unsupervised Learning. Clustering and the EM Algorithm. Unsupervised Learning is Model Learning

k-means Clustering David S. Rosenberg April 24, 2018 New York University

Unsupervised Learning and Clustering

Machine Learning. Unsupervised Learning. Manfred Huber

CS 8520: Artificial Intelligence. Machine Learning 2. Paula Matuszek Fall, CSC 8520 Fall Paula Matuszek

Artificial Intelligence. Programming Styles

CSE 573: Artificial Intelligence Autumn 2010

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

CS 8520: Artificial Intelligence

Weka ( )

Lecture 7: Clustering

10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors

Using Machine Learning to Optimize Storage Systems

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample

Evaluation. Evaluate what? For really large amounts of data... A: Use a validation set.

Unsupervised Learning : Clustering

Cluster Analysis. CSE634 Data Mining

UNSUPERVISED STATIC DISCRETIZATION METHODS IN DATA MINING. Daniela Joiţa Titu Maiorescu University, Bucharest, Romania

Introduction to Machine Learning CMU-10701

Unsupervised Learning. Supervised learning vs. unsupervised learning. What is Cluster Analysis? Applications of Cluster Analysis

ECLT 5810 Clustering

Introduction to Artificial Intelligence

Unsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing

Artificial Neural Networks (Feedforward Nets)

Association Rule Mining and Clustering

Mining di Dati Web. Lezione 3 - Clustering and Classification

Transcription:

Introduction to Information Retrieval http://informationretrieval.org IIR 16: Flat Clustering Hinrich Schütze Institute for Natural Language Processing, Universität Stuttgart 2009.06.16 Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters? 1/64 3/64 Overview 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters? MI eample for poultry/eport in Reuters ec = epoultry =1 ec = epoultry =0 et = eeport = 1 N11 = 49 N10 = 27,652 et = eeport = 0 N01 = 141 N00 = 774,106 these values into formula: Plug I (U; C) = 49 801,948 log 2 + 141 801,948 log 2 + 27,652 801,948 log 2 + 774,106 801,948 log 2 0.000105 801,948 49 (49+27,652)(49+141) 801,948 141 (141+774,106)(49+141) 801,948 27,652 (49+27,652)(27,652+774,106) 801,948 774,106 (141+774,106)(27,652+774,106)

Linear classifiers Linear classifiers compute a linear combination or weighted sum i w ii of the feature values. Classification decision: i w ii >θ? Geometrically, the equation i w ii = θ defines a line (2D), a plane (3D) or a hyperplane (higher dimensionalities). Assumption: The classes are linearly separable. Methods for finding a linear separator: Perceptron, Rocchio, Naive Bayes, many others A linear classifier in 2D A linear classifier in 2D is a line described by the equation w1d1 + w2d2 = θ Eample for a 2D linear classifier Points (d1 d2) with w1d1 + w2d2 θ are in the class c. Points (d1 d2) with w1d1 + w2d2 <θare in the complement class c. 5/64 7/64 A linear classifier in 1D A linear classifier in 3D A linear classifier in 1D a point described by th equation w1d1 = θ The point at θ/w1 Points (d1) with w1d1 are in the class c. Points (d1) with w1d1 < are in the complement class c. A linear classifier in 3D a plane described by th equation w1d1 + w2d2 + w3d3 = Eample for a 3D linea classifier Points (d1 d2 d3) with w1d1 + w2d2 + w3d3 are in the class c. Points (d1 d2 d3) with w1d1 + w2d2 + w3d3 < are in the complement class c.

Rocchio as a linear classifier Rocchio is a linear classifier defined by: M i=1 widi = w d = θ where the normal vector w = µ(c1) µ(c2) and θ =0.5 ( µ(c1) 2 µ(c2) 2 ). 9/64 knn is not a linear classifier The decision boundaries between classes are piecewise linear......but they are not linear separators that can be described as M i=1 w i di = θ. 11/64 Naive Bayes as a linear classifier Naive Bayes is a linear classifier defined by: M i=1 widi = θ where wi = log[ˆp(ti c)/ ˆP(ti c)], di = number of occurrences of ti in d, andθ = log[ˆp(c)/ˆp( c)]. Here, the inde i, 1 i M, refers to terms of the vocabulary (not to positions in d as k did in our original definition of Naive Bayes) Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters?

What is clustering? (Document) clustering is the process of grouping a set of documents into clusters of similar documents. Documents within a cluster should be similar. Documents from different clusters should be dissimilar. Clustering is the most common form of unsupervised learning. Unsupervised = there are no labeled or annotated data. 13/64 Data set with clear cluster structure 0.0 0.5 1.0 1.5 2.0 2.5 Classification vs. Clustering Classification: supervised learning Clustering: unsupervised learning Classification: Classes are human-defined and part of the input to the learning algorithm. Clustering: Clusters are inferred from the data without human input. However, there are many ways of influencing the outcome of clustering: number of clusters, similarity measure, representation of documents,... 15/64 0.0 0.5 1.0 1.5 2.0 Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters?

The cluster hypothesis Cluster hypothesis. Documents in the same cluster behave similarly with respect to relevance to information needs. All applications in IR are based (directly or indirectly) on the cluster hypothesis. Search result clustering for better navigation 17/64 19/64 Applications of clustering in IR Application What is Benefit Eample clustered? Search result clustering search results Scatter-Gather (subsets of) collection more effective information presentation to user alternative user interface: search without typing Collection clustering collection effective information presentation for eploratory browsing McKeown et al. 200 news.google.com Cluster-based retrieval collection higher efficiency: faster search Salton 1971 Scatter-Gather

Global navigation: MESH (lower level) Global navigation: Yahoo 23 / 64 21 / 64 Cartia Themescapes Google News But they are well known eamples for using a global hierarch for navigation. Some eamples for global navigation/eploration based on clustering: Note: Yahoo/MESH are not eamples of clustering. Global navigation: MESH (upper level)

Global navigation combined with visualization (1) Global clustering for navigation: Google News http://news.google.com 25/64 27/64 Global navigation combined with visualization (2) Clustering for improving recall To improve search recall: Cluster docs in collection a priori When a query matches a doc d, also return other docs in the cluster containing d Hope: if we do this: the query car will also return docs containing automobile Because clustering groups together docs containing car with those containing automobile. Both types of documents contain words like parts, dealer, mercedes, road trip.

Data set with clear cluster structure Eercise: Come up with an algorithm for finding the three clusters in this case 0.0 0.5 1.0 1.5 2.0 2.5 0.0 0.5 1.0 1.5 2.0 Issues in clustering General goal: put related docs in the same cluster, put unrelated docs in different clusters. But how do we formalize this? How many clusters? Initially, we will assume the number of clusters K is given. Often: secondary goals in clustering Eample: avoid very small and very large clusters Flat vs. hierarchical clustering Hard vs. soft clustering 29/64 31/64 Document representations in clustering Vector space model As in vector space classification, we measure relatedness between vectors by Euclidean distance...... which is almost equivalent to cosine similarity. Almost: centroids are not length-normalized. For centroids, distance and cosine give different results. Flat vs. Hierarchical clustering Flat algorithms Usually start with a random (partial) partitioning of docs into groups Refine iteratively Main algorithm: K -means Hierarchical algorithms Create a hierarchy Bottom-up, agglomerative Top-down, divisive

Hard vs. Soft clustering Hard clustering: Each document belongs to eactly one cluster. More common and easier to do Soft clustering: A document can belong to more than one cluster. Makes more sense for applications like creating browsable hierarchies You may want to put a pair of sneakers in two clusters: sports apparel shoes You can only do that with a soft clustering approach. We won t have time for soft clustering. See IIR 16.5, IIR 18 Flat algorithms Flat algorithms compute a partition of N documents into a set of K clusters. Given: a set of documents and the number K Find: a partition in K clusters that optimizes the chosen partitioning criterion Global optimization: ehaustively enumerate partitions, pick optimal one Not tractable Effective heuristic method: K-means algorithm 33/64 35/64 Our plan This lecture: Flat, hard clustering Net lecture: Hierarchical, hard clustering Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters?

K-means Perhaps the best known clustering algorithm Simple, works well in many cases Use as default / baseline for clustering documents K-means algorithm K-means({ 1,..., N}, K) 1 ( s1, s2,..., sk) SelectRandomSeeds({ 1,..., N}, K) 2 for k 1 to K 3 do µk sk 4 while stopping criterion has not been met 5 do for k 1 to K 6 do ωk {} 7 for n 1 to N 8 do j arg min j µj n 9 ωj ωj { n} (reassignment of vectors) 10 for k 1 to K 11 do µk 1 ωk 12 return { µ1,..., µk} ωk (recomputation of centroids) 37/64 39/64 K-means Each cluster in K-means is defined by a centroid. Objective/partitioning criterion: minimize the average squared difference from the centroid Recall definition of centroid: µ(ω) = 1 ω ω where we use ω to denote a cluster. We try to find the minimum average squared difference by iterating two steps: reassignment: assign each vector to its closest centroid recomputation: recompute each centroid as the average of the vectors that were assigned to it in reassignment K -means eample

K-means is guaranteed to converge Proof: The sum of squared distances (RSS) decreases during reassignment. RSS = sum of all squared distances between document vector and closest centroid (because each vector is moved to a closer centroid) RSS decreases during recomputation. (We will show this on the net slide.) There is only a finite number of clusterings. Thus: We must reach a fied point. (assume that ties are broken consistently) K-means is guaranteed to converge But we don t know how long convergence will take! If we don t care about a few docs switching back and forth, then convergence is usually fast (< 10-20 iterations). However, complete convergence can take many more iterations. 41/64 43/64 Recomputation decreases average distance RSS = K k=1 RSS k the residual sum of squares (the goodness measure) RSSk( v) = v 2 = M (vm m) 2 RSSk( v) vm ωk = ωk 2(vm m) =0 ωk m=1 vm = 1 ωk ωk m The last line is the componentwise definition of the centroid! We minimize RSSk when the old centroid is replaced with the new centroid. RSS, the sum of the RSSk, must then also decrease during recomputation. Optimality of K-means Convergence does not mean that we converge to the optimal clustering! This is the great weakness of K-means. If we start with a bad set of seeds, the resulting clustering can be horrible.

Eercise: Suboptimal clustering 3 2 d1 d2 d3 1 0 d4 d5 d6 0 1 2 3 4 What is the optimal clustering for K =2? Do we converge on this clustering for arbitrary seeds di1, d i2? Time compleity of K-means Computing one distance of two vectors is O(M). Reassignment step: O(KNM) (we need to compute KN document-centroid distances) Recomputation step: O(NM) (we need to add each of the document s < M values to one of the centroids) Assume number of iterations bounded by I Overall compleity: O(IKNM) linear in all important dimensions However: This is not a real worst-case analysis. In pathological cases, the number of iterations can be much higher than linear in the number of documents. 45/64 47/64 Initialization of K-means Random seed selection is just one of many ways K-means can be initialized. Random seed selection is not very robust: It s easy to get a suboptimal clustering. Better heuristics: Select seeds not randomly, but using some heuristic (e.g., filte outoutliersorfindasetofseedsthathas goodcoverage of the document space) Use hierarchical clustering to find good seeds (net class) Select i (e.g., i = 10) different sets of seeds, do a K -means clustering for each, select the clustering with lowest RSS Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters?

What is a good clustering? Internal criteria Eample of an internal criterion: RSS in K -means But an internal criterion often does not evaluate the actual utility of a clustering in the application. Alternative: Eternal criteria Evaluate with respect to a human-defined classification Eternal criterion: Purity purity(ω, C)= 1 N k ma j ωk cj Ω={ω1,ω2,...,ωK} is the set of clusters and C = {c1, c2,...,cj} is the set of classes. For each cluster ωk: find class cj with most members nkj in ωk Sum all nkj and divide by total number of points 49/64 51/64 Eternal criteria for clustering quality Based on a gold standard data set, e.g., the Reuters collection we also used for the evaluation of classification Goal: Clustering should reproduce the classes in the gold standard (But we only want to reproduce how documents are divided into groups, not the class labels.) First measure for how well we were able to reproduce the classes: purity Eample for computing purity cluster 1 cluster 2 cluster 3 o o o o o To compute purity: 5 = maj ω1 cj (class, cluster 1); 4 = maj ω2 cj (class o, cluster 2); and 3 = maj ω3 cj (class, cluster3). Purity is (1/17) (5 + 4 + 3) 0.71.

Rand inde Definition: RI = TP+TN TP+FP+FN+TN Based on 22 contingency table of all pairs of documents: same cluster different clusters same class true positives (TP) false negatives (FN) different classes false positives (FP) true negatives (TN) TP+FN+FP+TN is the total number of pairs. There are ( N 2) pairs for N documents. Eample: ( 17 ) 2 = 136 in o/ / eample Each pair is either positive or negative (the clustering puts the two documents in the same or in different clusters)...... and either true (correct) or false (incorrect): the clustering decision is correct or incorrect. Rand measure for the o/ / eample same cluster different clusters same class TP = 20 FN = 24 different classes FP = 20 TN = 72 RI is then (20 + 72)/(20 + 20 + 24 + 72) 0.68. 53/64 55/64 As an eample, we compute RI for the o/ / eample. We first compute TP + FP. The three clusters contain 6, 6, and 5 points, respectively, so the total number of positives or pairs of documents that are in the same cluster is: TP + FP = ( 6 2 ) + ( 6 2 ) + ( 5 2 ) =40 Of these, the pairs in cluster 1, the o pairs in cluster 2, the pairs in cluster 3, and the pair in cluster 3 are true positives: TP = ( 5 2 ) + ( 4 2 ) + ( 3 2 ) + ( 2 2 ) =20 Thus, FP = 40 20 = 20. FN and TN are computed similarly. Two other eternal evaluation measures Two other measures Normalized mutual information (NMI) How much information does the clustering contain about the classification? Singleton clusters (number of clusters = number of docs) have maimum MI Therefore: normalize by entropy of clusters and classes Fmeasure Like Rand, but precision and recall can be weighted

Evaluation results for the o/ / eample purity NMI RI F5 lower bound 0.0 0.0 0.0 0.0 maimum 1.0 1.0 1.0 1.0 value for eample 0.71 0.36 0.68 0.46 All four measures range from 0 (really bad clustering) to 1 (perfect clustering). How many clusters? Either: Number of clusters K is given. Then partition into K clusters K might be given because there is some eternal constraint. Eample: In the case of Scatter-Gather, it was hard to show more than 10 20 clusters on a monitor in the 90s. Or: Finding the right number of clusters is part of the problem. Given docs, find K for which an optimum is reached. How to define optimum? We can t use RSS or average squared distance from centroid as criterion: always chooses K = N clusters. 57/64 59/64 Outline 1 Recap 2 Clustering: Introduction 3 Clustering in IR 4 K-means 5 Evaluation 6 How many clusters? Eercise Suppose we want to analyze the set of all articles published by a major newspaper (e.g., New York Times or Süddeutsche Zeitung) in 2008. Goal: write a two-page report about what the major news stories in 2008 were. We want to use K-means clustering to find the major news stories. How would you determine K?

Simple objective function for K (1) Basic idea: Start with 1 cluster (K =1) Keep adding clusters (= keep increasing K ) Add a penalty for each new cluster Trade off cluster penalties against average squared distance from centroid Choose K with best tradeoff Finding the knee in the curve 2 4 6 8 10 61/64 Simple objective function for K (2) Given a clustering, define the cost for a document as (squared) distance to centroid Define total distortion RSS(K) as sum of all individual document costs (corresponds to average distance) Then: penalize each cluster with a cost λ Thus for a clustering with K clusters, total cluster penalty is Kλ Define the total cost of a clustering as distortion plus total cluster penalty: RSS(K) + Kλ Select K that minimizes (RSS(K) + Kλ) Still need to determine good value for λ... residual sum of squares 1750 1800 1850 1900 1950 number of clusters Pick the number of clusters where curve flattens. Here: 4 or 9. 63/64 Resources Chapter 16 of IIR Resources at http://ifnlp.org/ir K -means eample Keith van Rijsbergen on the cluster hypothesis (he was one of the originators) Bing/Carrot2/Clusty: search result clustering