Unsupervised Learning. Pantelis P. Analytis. Introduction. Finding structure in graphs. Clustering analysis. Dimensionality reduction.

Similar documents
Working with Unlabeled Data Clustering Analysis. Hsiao-Lung Chan Dept Electrical Engineering Chang Gung University, Taiwan

Data Mining and Data Warehousing Henryk Maciejewski Data Mining Clustering

Clustering Results. Result List Example. Clustering Results. Information Retrieval

Cluster Analysis: Agglomerate Hierarchical Clustering

Unsupervised Learning

Introduction to Data Mining

Big Data Analytics CSCI 4030

CHAPTER 4: CLUSTER ANALYSIS

Clustering. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

Part 1: Link Analysis & Page Rank

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

CS224W: Social and Information Network Analysis Jure Leskovec, Stanford University

Machine Learning and Data Mining. Clustering. (adapted from) Prof. Alexander Ihler

Unsupervised Learning and Data Mining

How to organize the Web?

Administrative. Machine learning code. Supervised learning (e.g. classification) Machine learning: Unsupervised learning" BANANAS APPLES

Clustering and Dimensionality Reduction. Stony Brook University CSE545, Fall 2017

4. Cluster Analysis. Francesc J. Ferri. Dept. d Informàtica. Universitat de València. Febrer F.J. Ferri (Univ. València) AIRF 2/ / 1

4. Ad-hoc I: Hierarchical clustering

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

K-Means Clustering 3/3/17

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Unsupervised Learning : Clustering

Hierarchical Clustering 4/5/17

Unsupervised Learning Hierarchical Methods

Clustering CS 550: Machine Learning

Lecture Notes to Big Data Management and Analytics Winter Term 2017/2018 Node Importance and Neighborhoods

Lecture-17: Clustering with K-Means (Contd: DT + Random Forest)

Clustering algorithms

CSE494 Information Retrieval Project C Report

Machine Learning (BSMC-GA 4439) Wenke Liu

Gene Clustering & Classification

Introduction to Artificial Intelligence

Cluster Analysis for Microarray Data

Unsupervised Learning

Supervised vs.unsupervised Learning

Lesson 3. Prof. Enza Messina

Unsupervised Learning and Clustering

ECG782: Multidimensional Digital Signal Processing

Methods for Intelligent Systems

Clustering Color/Intensity. Group together pixels of similar color/intensity.

Figure (5) Kohonen Self-Organized Map

DS504/CS586: Big Data Analytics Big Data Clustering Prof. Yanhua Li

Unsupervised Learning and Clustering

Data Informatics. Seon Ho Kim, Ph.D.

Unsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing

Statistics 202: Data Mining. c Jonathan Taylor. Clustering Based in part on slides from textbook, slides of Susan Holmes.

Measure of Distance. We wish to define the distance between two objects Distance metric between points:

Data Warehousing and Machine Learning

5/15/16. Computational Methods for Data Analysis. Massimo Poesio UNSUPERVISED LEARNING. Clustering. Unsupervised learning introduction

Uninformed Search Methods. Informed Search Methods. Midterm Exam 3/13/18. Thursday, March 15, 7:30 9:30 p.m. room 125 Ag Hall

Clustering Algorithm (DBSCAN) VISHAL BHARTI Computer Science Dept. GC, CUNY

Clustering. Unsupervised Learning

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Today s lecture. Clustering and unsupervised learning. Hierarchical clustering. K-means, K-medoids, VQ

CSE 5243 INTRO. TO DATA MINING

Preprocessing DWML, /33

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

Machine Learning (BSMC-GA 4439) Wenke Liu

Big Data Analytics CSCI 4030

Olmo S. Zavala Romero. Clustering Hierarchical Distance Group Dist. K-means. Center of Atmospheric Sciences, UNAM.

Clustering analysis of gene expression data

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

Introduction to Machine Learning. Xiaojin Zhu

Machine Learning. Unsupervised Learning. Manfred Huber

Clustering: K-means and Kernel K-means

Clustering. Partition unlabeled examples into disjoint subsets of clusters, such that:

Clustering k-mean clustering

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010

INF4820. Clustering. Erik Velldal. Nov. 17, University of Oslo. Erik Velldal INF / 22

Clustering. RNA-seq: What is it good for? Finding Similarly Expressed Genes. Data... And Lots of It!

A Tutorial on the Probabilistic Framework for Centrality Analysis, Community Detection and Clustering

CS325 Artificial Intelligence Ch. 20 Unsupervised Machine Learning

Clustering. Unsupervised Learning

Unsupervised Learning

STATS306B STATS306B. Clustering. Jonathan Taylor Department of Statistics Stanford University. June 3, 2010

Unsupervised learning on Color Images

Based on Raymond J. Mooney s slides

An Unsupervised Approach for Combining Scores of Outlier Detection Techniques, Based on Similarity Measures

Chapter 9. Classification and Clustering

CSE 40171: Artificial Intelligence. Learning from Data: Unsupervised Learning

Information Networks: PageRank

Unsupervised: no target value to predict

Exploratory Data Analysis using Self-Organizing Maps. Madhumanti Ray

Computational Statistics The basics of maximum likelihood estimation, Bayesian estimation, object recognitions

Midterm Examination CS540-2: Introduction to Artificial Intelligence

CS249: SPECIAL TOPICS MINING INFORMATION/SOCIAL NETWORKS

Network Traffic Measurements and Analysis

Hierarchical clustering

K Nearest Neighbor Wrap Up K- Means Clustering. Slides adapted from Prof. Carpuat

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

Applications. Foreground / background segmentation Finding skin-colored regions. Finding the moving objects. Intelligent scissors

Mining Web Data. Lijun Zhang

Flat Clustering. Slides are mostly from Hinrich Schütze. March 27, 2017

Clustering in Data Mining

Finding Clusters 1 / 60

MIDTERM EXAMINATION Networked Life (NETS 112) November 21, 2013 Prof. Michael Kearns

Transcription:

March 19, 2018 1 / 40

1 2 3 4 2 / 40

What s unsupervised learning? Most of the data available on the internet do not have labels. How can we make sense of it? 3 / 40

4 / 40

5 / 40

Organizing the web First attempts to organize the web were based on human curated directories (Yahoo, looksmart). People also used methods from information retrieval to uncover relevant documents. Yet he web has a deluge of untrusted documents, spam, random webpages, advertisements etc. 6 / 40

Elements of the PageRank algorithm Solution: Use social feedback to rank the quality of documents. You can see links as vote. A page is more important when it has more incoming links. For instance www.nytimes.com has numerous incoming notes, as opposed to www.inkefalonia.gr Links from important questions count more Recursiveness. 7 / 40

The iterative PageRank algorithm At t = 0, assume an initial probability distribution: PR(p i ; 0) = 1 N. At each time step, the computation yields: PR(p i ; t + 1) = 1 d N + d p j M(p i ) PR(p j ;t) L(p j ) 8 / 40

At t = 0, assume an initial probability distribution: PR(p i ; 0) = 1 N. At each time step, the computation yields: PR(p i ; t + 1) = 1 d N + d PR(p j ;t) p j M(p i ) L(p j ) 9 / 40

10 / 40

Page Rank Equilibrium 11 / 40

PageRank: The spider trap 12 / 40

PageRank: The spider trap 13 / 40

The Scaled PageRank algorithm Scaled PageRank Update Rule Apply basic PR rule. Scale all values down by factor s. Divide the 1-s leftover units of PR evenly over nodes. 14 / 40

What s unsupervised learning? Most of the data available on the internet do not have labels. How can we make sense of it? 15 / 40

: the k-means algorithm Input: K, set of points x 1,..., x n Place centroids c 1,..., c k randomly Then repeat until convergence: For each point x i find the nearest centroid c j and assign that point to that cluster In math notation: argmin j D(x i, c j ) For each cluster j = 1,..., K find the new centroid of all points x i assigned to cluster j in previous step. In math notation: c j (a) = 1/n j x i (a) for a = 1,..., d x i c j Stop when the algorithm has converged i.e. none of the items changes cluster. 16 / 40

Converging to clusters 17 / 40

How do we select k? There are diminishing returns in the size of different clusters. An intuitive approach suggests picking the after which the distance flattens out. 18 / 40

Hierarchical 19 / 40

Agglomerative vs. divisive Agglomerative clustering starts from the bottom and moves to larger clusters. Divisive clustering starts with one cluster which is gradually disintegrated into smaller ones. 20 / 40

Agglomerative vs. divisive How do we determine the nearness of clusters? Complete linkage: D(X, Y ) = max x X,y Y d(x, y) Single linkage: D(X, Y ) = min x X,y Y d(x, y) Average linkage: 1 X Y x X y Y d(x, y). 21 / 40

Agglomerative Pick k upfront, stop when we have k clusters. Stop when a cluster with low cohesion is created (diameter, radius or density-based approaches). 22 / 40

Kohonen s self-organizing maps Step 0: Randomly position the grid s neurons in the data space. 23 / 40

Kohonen s self-organizing maps Step 1: Select one data point, either randomly or systematically cycling through the dataset in order 24 / 40

Kohonen s self-organizing maps Step 2: Find the neuron that is closest to the chosen data point. This neuron is called the Best Matching Unit (BMU). 25 / 40

Kohonen s self-organizing maps Step 3: Move the BMU closer to that data point. The distance moved by the BMU is determined by a learning rate, which decreases after each iteration. 26 / 40

Kohonen s self-organizing maps Step 4: Move the BMU s neighbors closer to that data point as well, with farther away neighbors moving less. Neighbors are identified using a radius around the BMU, and the value for this radius decreases after each iteration. 27 / 40

Kohonen s self-organizing maps 28 / 40

Kohonen s self-organizing maps 29 / 40

Kohonen s self-organizing maps 30 / 40

Kohonen s self-organizing maps 31 / 40

Kohonen s self-organizing maps Update the learning rate and BMU radius, before repeating Steps 1 to 4. Iterate these steps until positions of neurons have been stabilized. 32 / 40

Kohonen s self-organizing maps 33 / 40

Principal component 34 / 40

Principal component 35 / 40

Principal component 36 / 40

Principal component Often used to accelerate supervised learning. Visualization 37 / 40

Principal component 38 / 40

Principal component 39 / 40

in recommender systems 40 / 40