Machine Learning : Clustering, Self-Organizing Maps

Similar documents
Image Processing. Image Features

Neural networks for variable star classification

Unsupervised Learning

Figure (5) Kohonen Self-Organized Map

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

Unsupervised learning in Vision

ECG782: Multidimensional Digital Signal Processing

Function approximation using RBF network. 10 basis functions and 25 data points.

K-Means Clustering. Sargur Srihari

Intro to Artificial Intelligence

Mixture Models and EM

Region-based Segmentation

Introduction to Machine Learning CMU-10701

Pattern recognition. Classification/Clustering GW Chapter 12 (some concepts) Textures

Pattern recognition. Classification/Clustering GW Chapter 12 (some concepts) Textures

9.1. K-means Clustering

Today. Gradient descent for minimization of functions of real variables. Multi-dimensional scaling. Self-organizing maps

Content-based Image and Video Retrieval. Image Segmentation

Machine learning Pattern recognition. Classification/Clustering GW Chapter 12 (some concepts) Textures

Artificial Neural Networks Unsupervised learning: SOM

Contents Machine Learning concepts 4 Learning Algorithm 4 Predictive Model (Model) 4 Model, Classification 4 Model, Regression 4 Representation

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

Using Machine Learning to Optimize Storage Systems

Unsupervised Learning

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2017

Clustering CS 550: Machine Learning

PATTERN RECOGNITION USING NEURAL NETWORKS

Classifying Images with Visual/Textual Cues. By Steven Kappes and Yan Cao

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

INF 4300 Classification III Anne Solberg The agenda today:

Based on Raymond J. Mooney s slides

Norbert Schuff VA Medical Center and UCSF

Neural Networks. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

The exam is closed book, closed notes except your one-page cheat sheet.

Two-step Modified SOM for Parallel Calculation

Action Recognition & Categories via Spatial-Temporal Features

Kapitel 4: Clustering

University of Florida CISE department Gator Engineering. Clustering Part 2

More on Learning. Neural Nets Support Vectors Machines Unsupervised Learning (Clustering) K-Means Expectation-Maximization

10-701/15-781, Fall 2006, Final

Robot Learning. There are generally three types of robot learning: Learning from data. Learning by demonstration. Reinforcement learning

COMS 4771 Clustering. Nakul Verma

Nonlinear dimensionality reduction of large datasets for data exploration

Associative Cellular Learning Automata and its Applications

What to come. There will be a few more topics we will cover on supervised learning

Clustering: Classic Methods and Modern Views

Global Probability of Boundary

Dimension Reduction CS534

Unsupervised Learning : Clustering

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation.

CHAPTER 3 TUMOR DETECTION BASED ON NEURO-FUZZY TECHNIQUE

CSE 5526: Introduction to Neural Networks Radial Basis Function (RBF) Networks

Unsupervised Learning

Introduction to Computer Science

Machine Learning A W 1sst KU. b) [1 P] Give an example for a probability distributions P (A, B, C) that disproves

Intelligent Image and Graphics Processing

Perceptron: This is convolution!

Applying Supervised Learning

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Voronoi Region. K-means method for Signal Compression: Vector Quantization. Compression Formula 11/20/2013

[7.3, EA], [9.1, CMB]

IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 6, NOVEMBER Inverting Feedforward Neural Networks Using Linear and Nonlinear Programming

1 Case study of SVM (Rob)

Clustering in Data Mining

Facial Expression Recognition Using Non-negative Matrix Factorization

Support Vector Machines

Modified Self-Organizing Mixture Network for Probability Density Estimation and Classification

Rough Set Approach to Unsupervised Neural Network based Pattern Classifier

A Dendrogram. Bioinformatics (Lec 17)

Exploratory data analysis for microarrays

Unsupervised Learning Partitioning Methods

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

Chapter 7: Competitive learning, clustering, and self-organizing maps

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

Simulation of WSN in NetSim Clustering using Self-Organizing Map Neural Network

Images Reconstruction using an iterative SOM based algorithm.

CAMCOS Report Day. December 9 th, 2015 San Jose State University Project Theme: Classification

Slide07 Haykin Chapter 9: Self-Organizing Maps

Supervised vs.unsupervised Learning

Introduction to Artificial Intelligence

Operators-Based on Second Derivative double derivative Laplacian operator Laplacian Operator Laplacian Of Gaussian (LOG) Operator LOG

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

On Classification: An Empirical Study of Existing Algorithms Based on Two Kaggle Competitions

The Traveling Salesman

Image enhancement for face recognition using color segmentation and Edge detection algorithm

Unsupervised Learning: Clustering

INTRODUCTION TO ARTIFICIAL INTELLIGENCE

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

A NEW ALGORITHM FOR OPTIMIZING THE SELF- ORGANIZING MAP

Neural Networks. Single-layer neural network. CSE 446: Machine Learning Emily Fox University of Washington March 10, /10/2017

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

CAD SYSTEM FOR AUTOMATIC DETECTION OF BRAIN TUMOR THROUGH MRI BRAIN TUMOR DETECTION USING HPACO CHAPTER V BRAIN TUMOR DETECTION USING HPACO

Assignment 2. Classification and Regression using Linear Networks, Multilayer Perceptron Networks, and Radial Basis Functions

Cluster analysis of 3D seismic data for oil and gas exploration

Automatic Photo Popup

Computer Vision Group Prof. Daniel Cremers. 6. Boosting

Multiple cosegmentation

A Keypoint Descriptor Inspired by Retinal Computation

Subspace Clustering. Weiwei Feng. December 11, 2015

Transcription:

Machine Learning Clustering, Self-Organizing Maps 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps

Clustering The task: partition a set of objects into meaningful subsets (clusters). The objects in a subset should be similar. Notations: Set of Clusters Set of indices Feature vectors Partitioning for, 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 2

Clustering Let and each cluster has a representative The task reads: Alternative variant is to consider the clustering as a mapping that assigns a cluster number to each 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 3

K-Means Algorithm Initialize centers randomly, Repeat until convergence: 1. Classify: 2. Update centers: The task is NP converges to a local optimum (depends on the initialization) 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 4

Sequential K-Means Repeat infinitely: 1. Chose randomly a feature vector from the training data 2. Classify it: 3. Update the -th center: with a decreasing step converges to the same, as the parallel version is a special case of Robbins-Monro Algorithm 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 5

Some variants Other distances, e.g. instead of In the K-Means algorithm the classification step remains the same, the update step the geometric median of (a bit complicated as the average ). Another problem: features may be not additive ( does not exist) Solution: K-Medioid Algorithm ( training set) is a feature vector from the 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 6

A generalization Observe (for the Euclidean distance): In what follows: with a Distance Matrix that can be defined in very different ways. Example: Objects are nodes of a weighted graph, of the shortest path from to. is the length Distances for other objects (non-vectors): Edit (Levenshtein) distance between two symbolic sequences For graphs distances based on graph isomorphism etc. 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 7

An application color reduction Objects are pixels, features are RGB-values. Partition the RGB-space into characteristic colors. (8 colors) 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 8

Another application superpixel segmentation Object are pixels. Features are RGBXY-values. Those pixels belong to the same cluster that are close to each other both spatially and in in the RGB-space SLIC Superpixels: http://ivrg.epfl.ch/research/superpixels 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 9

Cohonen Networks, Self-Organizing Maps The task is to approximate a dataset by a neural network of a certain topology. An example stereo in flatland. The input space is 3- (or more) dimensional, the set of points is however isomorphic to a 2D-space (up to noises). 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 10

Self-Organizing Maps SOM-s (usually) consist of RBF-neurons, each one represents (covers) a part of the input space (specified by the centers ). The network topology is given by means of a distance. Example neurons are nodes of a weighted graph, distances are shortest paths. For the flatland example the graph is a 2D-grid with unit weight for all edges. 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 11

Self-Organizing Maps, sequential algorithm 1. Chose randomly a feature vector from the training data (white) 2. Compute the winner -neuron (dark-yellow) 3. Compute the neighborhood of in the network (yellow) 4. Update the weights of all neurons from 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 12

Self-Organizing Maps, algorithms is monotonously decreasing with respect to (time) and Without 3) the sequential K-Means. Parallel variants: Go through all feature vectors, sum up the gradients, apply. Example for : The network fits into the data distribution (unfolds). 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 13

K-Means Expectation Maximization EM compute posteriors for the Expectation step K-Means classify object 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 14

Conclusion Before: 1. Probability theory models, inference, learning 2. Discriminative learning Classifiers Neural networks: 1. Feed-Forward Networks complex classifiers 2. Hopfield Networks structured output 3. Cohonen Networks clustering (unsupervised), model fitting Next topics further classifiers: 1. Support Vector Machines, Kernels 2. Empirical Risk minimization 3. Principal Component Analysis 4. Combining classifiers Decision Trees, AdaBoost 12/12/2013 Machine Learning : Clustering, Self-Organizing Maps 15