ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

Similar documents
SYDE Winter 2011 Introduction to Pattern Recognition. Clustering

9881 Project Deliverable #1

CSE 40171: Artificial Intelligence. Learning from Data: Unsupervised Learning

Clustering and Visualisation of Data

Cluster Analysis. Angela Montanari and Laura Anderlucci

Natural Language Processing

Machine Learning for Data Science (CS4786) Lecture 11

Clustering. RNA-seq: What is it good for? Finding Similarly Expressed Genes. Data... And Lots of It!

Homework Assignment #3

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

ADVANCED MACHINE LEARNING MACHINE LEARNING. Kernel for Clustering kernel K-Means

SGN (4 cr) Chapter 11

Clustering Lecture 3: Hierarchical Methods

Other conditional and loop constructs. Fundamentals of Computer Science Keith Vertanen

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

What is Clustering? Clustering. Characterizing Cluster Methods. Clusters. Cluster Validity. Basic Clustering Methodology

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

Image Segmentation. Srikumar Ramalingam School of Computing University of Utah. Slides borrowed from Ross Whitaker

8. Clustering: Pattern Classification by Distance Functions

Cluster Analysis. Summer School on Geocomputation. 27 June July 2011 Vysoké Pole

Segmentation Computer Vision Spring 2018, Lecture 27

Unsupervised Learning. Presenter: Anil Sharma, PhD Scholar, IIIT-Delhi

Clustering. So far in the course. Clustering. Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. dist(x, y) = x y 2 2

Processing and Others. Xiaojun Qi -- REU Site Program in CVMA

Enters system mode. Example The following example creates a scheduler named maintenancesched and commits the transaction:

Administration. Final Exam: Next Tuesday, 12/6 12:30, in class. HW 7: Due on Thursday 12/1. Final Projects:

Lecture 15 Clustering. Oct

Hierarchical Clustering 4/5/17

Clustering. Subhransu Maji. CMPSCI 689: Machine Learning. 2 April April 2015

Clustering algorithms and introduction to persistent homology

Unsupervised Learning

Today s lecture. Clustering and unsupervised learning. Hierarchical clustering. K-means, K-medoids, VQ

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

Text Documents clustering using K Means Algorithm

Lecture 8: Fitting. Tuesday, Sept 25

Physics 2210 Fall smartphysics 02 Kinematics in 2- and 3-d 08/31/2015

Nights & Weekend ROB & PHIL DECEMBER 12, 2008

Pattern Clustering with Similarity Measures

Stat 321: Transposable Data Clustering

Expectation Maximization!

Pattern Recognition Lecture Sequential Clustering

Cluster Analysis: Agglomerate Hierarchical Clustering

APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE

INF4820, Algorithms for AI and NLP: Hierarchical Clustering

Document Clustering: Comparison of Similarity Measures

Applications. Foreground / background segmentation Finding skin-colored regions. Finding the moving objects. Intelligent scissors

COSC 6397 Big Data Analytics. Fuzzy Clustering. Some slides based on a lecture by Prof. Shishir Shah. Edgar Gabriel Spring 2015.

Fuzzy Segmentation. Chapter Introduction. 4.2 Unsupervised Clustering.

Cluster Analysis. Ying Shen, SSE, Tongji University

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/09/2018)

12 Dynamic Programming (2) Matrix-chain Multiplication Segmented Least Squares

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Machine learning - HT Clustering

INF4820 Algorithms for AI and NLP. Evaluating Classifiers Clustering

Greedy Algorithms. Algorithms

Behavioral Data Mining. Lecture 18 Clustering

Digital Geometry Processing

View a Students Schedule Through Student Services Trigger:

CLUSTERING ALGORITHMS

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Problem Definition. Clustering nonlinearly separable data:

Part I. Hierarchical clustering. Hierarchical Clustering. Hierarchical clustering. Produces a set of nested clusters organized as a

Notes. Reminder: HW2 Due Today by 11:59PM. Review session on Thursday. Midterm next Tuesday (10/10/2017)

61A Lecture 7. Monday, September 15

Clustering Algorithms for general similarity measures

Indexing in Search Engines based on Pipelining Architecture using Single Link HAC

CSL 860: Modern Parallel

Lecture 8 Object Descriptors

Outline. Advanced Digital Image Processing and Others. Importance of Segmentation (Cont.) Importance of Segmentation

VK Multimedia Information Systems

Minoru SASAKI and Kenji KITA. Department of Information Science & Intelligent Systems. Faculty of Engineering, Tokushima University

Methods for Intelligent Systems

CSE 158. Web Mining and Recommender Systems. Midterm recap

CHAPTER 4: CLUSTER ANALYSIS

Clustering. Informal goal. General types of clustering. Applications: Clustering in information search and analysis. Example applications in search

CS 534: Computer Vision Segmentation and Perceptual Grouping

Introduction to Information Retrieval

CS 664 Slides #11 Image Segmentation. Prof. Dan Huttenlocher Fall 2003

6. Dicretization methods 6.1 The purpose of discretization

Chapter DM:II. II. Cluster Analysis

Spectral Classification

Machine Learning. Unsupervised Learning. Manfred Huber

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample

Clustering. Pattern Recognition IX. Michal Haindl. Clustering. Outline

Case-Based Reasoning. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. Parametric / Non-parametric.

CS 188: Artificial Intelligence Fall 2008

Clustering COMS 4771

Machine Learning. B. Unsupervised Learning B.1 Cluster Analysis. Lars Schmidt-Thieme

Chapter 6: Cluster Analysis

Types of general clustering methods. Clustering Algorithms for general similarity measures. Similarity between clusters

Data Mining Cluster Analysis: Basic Concepts and Algorithms. Lecture Notes for Chapter 8. Introduction to Data Mining

Machine Learning (BSMC-GA 4439) Wenke Liu

CS 231A CA Session: Problem Set 4 Review. Kevin Chen May 13, 2016

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

A Framework for Hierarchical Clustering Based Indexing in Search Engines

Hierarchical and Ensemble Clustering

Clustering. CS294 Practical Machine Learning Junming Yin 10/09/06

Unsupervised Learning Partitioning Methods

1 Case study of SVM (Rob)

Crystal Reports. Overview. Contents. Cross-Tab Capabilities & Limitations in Crystal Reports (CR) 6.x

Transcription:

Memorial University of Newfoundland Pattern Recognition Lecture 15, June 29, 2006 http://www.engr.mun.ca/~charlesr Office Hours: Tuesdays & Thursdays 8:30-9:30 PM EN-3026 July 2006 Sunday Monday Tuesday Wednesday Thursday Friday Saturday 2 4 6 8 Lecture 16 Lecture 17 Assignment 4 Due 9 11 13 15 Lecture 18 Lecture 19 16 Presentations 18 20 22 Presentations Assignment 5 Due 23 25 27 29 Lecture 21 Lecture 22 Final Reports 2

Last Week Clustering (Unsupervised Classification) Distribution Modes and Minima Pattern Grouping - Group similar patterns using distance metrics. - Merge or split clusters based on cluster similarity measurements. - Measures of clusters goodness # occurrences ß pel value 3 Recap: Simple Grouping using Threshold 1.! k = 1!! (number of clusters) 2.! z1 = x1!! (set first sample as class prototype) 3.! For all other samples xi:! a.! Find zj for which d(xi,zj) is minimum " b." If d(xi,zj) # T then assign xi to Cj! c.! Else k = k+1, zk = xi 4

Recap: Hierarchical Grouping Algorithms Successively merge based on some measure of within or between cluster similarity. For n samples {x1,..., xn} 1.! k = n, Ci = xi for i = 1...n 2.! Merge Ci, Cj which are most similar, k = k-1: min d(c i, C j ) C i = C i C j i,j 3.! Continue until some stopping condition is met. 5 Goodness of Partitioning Several of the stopping conditions suggest that we can use a measure of the scatter of each cluster to gauge how good the overall clustering is. In general, we would like compact clusters with a lot of space between them. We can use the measure of goodness to iteratively move samples from one cluster to another to optimize the groupings. 6

Clustering Criterion Global measurements of the goodness of the clusters. 1. Representative error = summed scatter within clusters Representation error of a clustering is the error from representing the N samples by the k cluster prototypes. Can choose zi to minimize J i = xɛc i x z i 2 7 So Now define scatter matrix Define summed scatter to be Thus J e = tr S W S W = = k i=1 S Wi 8

Clustering Criterion 2. Use volume of summed scatter: S w 9 Clustering Criterion 3. Could use between-cluster to within-cluster scatter So could use S B = k N i (m i m)(m i m) T i=1 tr Sw 1 S B Note: Any within-cluster criterion is minimized with k=n, and thus we would need an independent criterion for k. 10

K-Means Once we have a criterion, we can create an iterative clustering scheme. K-Means is the classic iterative clustering scheme: 1. Choose k prototypes {z1,..., zk} 2. Assign all samples to clusters: xɛc i if d(x, z i ) < d(x, z j ) j i 3. Update {zi} to minimize Ji, i=1,...,k { } z i = 1 x = m N i i 4. Reassign the samples using the new prototypes 5. Continue until no prototypes change. C i 11 ENG 8801/9881 From - KM Special Demo Topics Algorithm in Computer Java Engineering: Applet" Pattern " Recognition " http://web.sfc.keio.ac.jp/~osada/km/index.html 12

K-Means Good Points:! - Simple conceptually! - Successful if k is accurate and clusters are well-separated Problems: " - If k is incorrect, then the clusters can t be right! - Efficiency depends on sample order! - Non-spherical clusters cause problems 13 Extensions to K-Means There are several ways to extend the basic k-means algorithm: 1. Global minimization of Je! As an alternative to simply assigning samples to closest cluster prototype. 2. Allow a variable k 14

J e = k i=1 J i Basic Plan: Global Minimization of Je Ji is the representation error for the i th cluster. Move sample x from Ci to Cj if the magnitude of the increment to the representation error in Jj is less than the decrement to the representation error in Ji δ j < δ i 15 16

17 Similarly, we can show that the decrement to Ji is δ i = N i N i 1 x m i 2 So the reassignment rule (step 2 of K-means) is: - Move x from Ci to Cj if N j N j + 1 x m j 2 < N i N i 1 x m i 2 18

Notes on Global Minimization 1. Rule has little impact when Ni and Nj are very large 2. A point nearly on the MED boundary will be reassigned since N j while N j + 1 < 1 1 N i N i 1 > 1 no matter what Ni and Nj are. 3. If x is an unassigned sample, would get minimum increase to Je with: N i N i + 1 x m i 2 C j C i N j N j + 1 x m j 2 Modifies initial K-Means assignment by taking cluster size into account. If x is equidistant from mi, mj, then assign it to the smallest cluster. 19 Example Consider the following set of samples: (0,2) (1,0) (1,1)(1,2) (1,3) (1,4) (2,2) (3,1) (4,1) (5,0) (5,1) (5,2) (6,1) (7,1) Cluster using basic K-means and using K-means with global minimization method. Use k = 2. What happens if we start with k $ 2 (e.g. k = 3)? 20

Dealing with K Need a way of varying k in accordance with goodness of partitioning. Strategies for dealing with k: 1. Delete clusters with too few samples.! If Ni < T1 drop Ci and zi, and reassign samples from Ci. 2. Merge clusters which are close together.! If (m i m j ) T ( Si + S j 2 ) (m i m j ) < T 2! then replace Ci with the union of Ci and Cj, and drop zj. 3. Split clusters which are spread out.! If maximum eigenvalue of Si is greater than T3, split Ci with a plane through mi perpendicular to max. eigenvector and add new cluster. 21 There are many possible clustering algorithms. See Statistical Pattern Recognition: A Review by Jain, Duin, and Mao for more possibilities. 22