Exploratory Data Analysis using Self-Organizing Maps. Madhumanti Ray

Similar documents
CIE L*a*b* color model

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Chapter 7: Competitive learning, clustering, and self-organizing maps

arxiv: v1 [physics.data-an] 27 Sep 2007

10/14/2017. Dejan Sarka. Anomaly Detection. Sponsors

Unsupervised Data Mining: Clustering. Izabela Moise, Evangelos Pournaras, Dirk Helbing

Unsupervised Learning

CHAPTER FOUR NEURAL NETWORK SELF- ORGANIZING MAP

Clustering. Chapter 10 in Introduction to statistical learning

CS 1675 Introduction to Machine Learning Lecture 18. Clustering. Clustering. Groups together similar instances in the data sample

Hard clustering. Each object is assigned to one and only one cluster. Hierarchical clustering is usually hard. Soft (fuzzy) clustering

Network Traffic Measurements and Analysis

Data Preprocessing. Javier Béjar. URL - Spring 2018 CS - MAI 1/78 BY: $\

Artificial Neural Networks Unsupervised learning: SOM

Function approximation using RBF network. 10 basis functions and 25 data points.

Introduction to Artificial Intelligence

Topic 6 Representation and Description

Machine Learning Methods in Visualisation for Big Data 2018

Methods for Intelligent Systems

Clustering. Robert M. Haralick. Computer Science, Graduate Center City University of New York

Unsupervised Learning : Clustering

COMBINED METHOD TO VISUALISE AND REDUCE DIMENSIONALITY OF THE FINANCIAL DATA SETS

Clustering. Mihaela van der Schaar. January 27, Department of Engineering Science University of Oxford

Unsupervised Learning

Data Mining Chapter 9: Descriptive Modeling Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Machine Learning for OR & FE

Unsupervised Learning

9/17/2009. Wenyan Li (Emily Li) Sep. 15, Introduction to Clustering Analysis

Data Preprocessing. Javier Béjar AMLT /2017 CS - MAI. (CS - MAI) Data Preprocessing AMLT / / 71 BY: $\

Segmentation Computer Vision Spring 2018, Lecture 27

Forestry Applied Multivariate Statistics. Cluster Analysis

ECLT 5810 Clustering

University of Florida CISE department Gator Engineering. Clustering Part 2

Unsupervised Learning and Clustering

ADVANCED MACHINE LEARNING MACHINE LEARNING. Kernel for Clustering kernel K-Means

ECLT 5810 Clustering

Data Warehousing and Machine Learning

Clustering algorithms

Intelligent Agent Technology in Modern Production and Trade Management

k-means Clustering Todd W. Neller Gettysburg College

Clustering. Supervised vs. Unsupervised Learning

Data Mining and Data Warehousing Henryk Maciejewski Data Mining Clustering

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Exploratory data analysis for microarrays

A Dendrogram. Bioinformatics (Lec 17)

Non-Bayesian Classifiers Part I: k-nearest Neighbor Classifier and Distance Functions

ECG782: Multidimensional Digital Signal Processing

Preprocessing DWML, /33

An Introduction to Cluster Analysis. Zhaoxia Yu Department of Statistics Vice Chair of Undergraduate Affairs

Automatic Group-Outlier Detection

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Chapter 6: Cluster Analysis

University of Florida CISE department Gator Engineering. Clustering Part 4

IBL and clustering. Relationship of IBL with CBR

CHAPTER 4: CLUSTER ANALYSIS

Multivariate Analysis (slides 9)

Clustering COMS 4771

Clustering & Classification (chapter 15)

Cluster Analysis. Prof. Thomas B. Fomby Department of Economics Southern Methodist University Dallas, TX April 2008 April 2010

CLUSTERING. CSE 634 Data Mining Prof. Anita Wasilewska TEAM 16

Clustering Part 4 DBSCAN

Clustering and Visualisation of Data

Unsupervised Learning and Clustering

Hierarchical Clustering 4/5/17

APPLICATION OF MULTIPLE RANDOM CENTROID (MRC) BASED K-MEANS CLUSTERING ALGORITHM IN INSURANCE A REVIEW ARTICLE

INF4820, Algorithms for AI and NLP: Evaluating Classifiers Clustering

Based on Raymond J. Mooney s slides

Gene Clustering & Classification

Unsupervised Learning

Unsupervised learning

Lecture 3 January 22

Supervised vs. Unsupervised Learning

5.6 Self-organizing maps (SOM) [Book, Sect. 10.3]

Dimension reduction : PCA and Clustering

What to come. There will be a few more topics we will cover on supervised learning

Understanding Clustering Supervising the unsupervised

Case-Based Reasoning. CS 188: Artificial Intelligence Fall Nearest-Neighbor Classification. Parametric / Non-parametric.

CS 188: Artificial Intelligence Fall 2008

L9: Hierarchical Clustering

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

Data Mining: Data. Lecture Notes for Chapter 2. Introduction to Data Mining

K-means Clustering of Proportional Data Using L1 Distance

Supervised vs.unsupervised Learning

COMP 551 Applied Machine Learning Lecture 13: Unsupervised learning

Kernels and Clustering

Preface to the Second Edition. Preface to the First Edition. 1 Introduction 1

CPSC 340: Machine Learning and Data Mining. Multi-Dimensional Scaling Fall 2017

Color Image Segmentation Using a Spatial K-Means Clustering Algorithm

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

Cluster Analysis: Agglomerate Hierarchical Clustering

Giri Narasimhan. CAP 5510: Introduction to Bioinformatics. ECS 254; Phone: x3748

Figure (5) Kohonen Self-Organized Map

CS 2750 Machine Learning. Lecture 19. Clustering. CS 2750 Machine Learning. Clustering. Groups together similar instances in the data sample

INF 4300 Classification III Anne Solberg The agenda today:

MSA220 - Statistical Learning for Big Data

Table Of Contents: xix Foreword to Second Edition

CSE 40171: Artificial Intelligence. Learning from Data: Unsupervised Learning

Clustering Algorithms. Margareta Ackerman

Unsupervised Learning Partitioning Methods

k-means Clustering Todd W. Neller Gettysburg College Laura E. Brown Michigan Technological University

Transcription:

Exploratory Data Analysis using Self-Organizing Maps Madhumanti Ray

Content Introduction Data Analysis methods Self-Organizing Maps Conclusion

Visualization of high-dimensional data items Exploratory data analysis finds structures in a given data set. Many methods exist deal with low dimensional data Eg: Andrews curve, Chernoff' faces Drawback? Difficult to visualize high-dimensional data, do not reduce amount of data

Data Exploration Techniques Clustering Projection

Clustering Reduces amount of data by grouping similar items together. Hierarchical Partitioning Partitioning : divide the data set into disjoint data clusters

K- means Algorithm Find Euclidean distance between each input and centroid of a cluster Assign input to nearest cluster Do above steps iteratively till value does not change significantly

Projection Methods Represents high-dimensional data in a lower-dimensional space Properties of the structure of data set is preserved Linear Projection methods Non-linear projection methods

Linear Projection Principal component analysis: Data sets with linear structures Original variables mapped to few variable, PC, retain key features of data set Projection Pursuit: Data projection deviating most from normally distribute data is picked for removal

Non-linear Projection methods Multidimensional Scaling: Topology preserving Distance between representations in lowerdimensional space close to original distance Objective function: Σ [d(k,l) d'(k,l)] 2 Rank order of distances maintained for comparison maintained by non-metric MDS better measure

Sammon's : MDS function normalized by original distance Principal curves: points on curve is average of all data points projecting to it.

Self Organizing Maps Unsupervised learning, non-parametric Can do both clustering and projection Preserves topology Implements competitive learning with cooperation

Self Organizing Maps Winning unit on lattice has reference vector closest to input: c = c(x) = arg min{ x-m i 2 } euclidean distance to find winning unit

Self Organizing Maps The learning is shared by neighbouring units - they orient their vectors towards input Governed by neighbourhood kernel h decreasing function m i (t + 1) = m i (t) + h ci (t)[x(t) m i (t)] h is wide at first, reduces over learning

Features of SOM useful in exploring data Ordered display: neighbouring units on the map will represent inputs that are similar Visualization of clusters: distance between reference vectors gives density of cluster used for gray level values. Missing data : reference vectors compared to input using features present in input Outliers: affect only one unit and neighbours.

Using SOM Preprocessing: Feature extraction that represent the variations of data best. Ordering/scaling among feature components. Computation of maps: Initialize reference vectors, choose wide h Neighbourhood kernel h reduces in phase one Phase two, reaches final state of map

Using SOM Choosing best map: Teach set of maps with each size of maps and h Pick map that minimizes cost function best from each set Apply goodness measure for final pick Interpretation of map: Subjective to data set and problem SOM maps focused on local relations of data items Evaluation: Many methods, eg. Classification accuracy Generalizability

Difference with other MDS With MDS: Topology preserving, not distance i.e. similar inputs mapped to nearby locations Computational complexity depends on map units k. Each learning step is O(k) computations No local minima. Start with wide h, shrink to narrow form

Conclusion SOM: Viable alternative to other data exploration methods One method to implement clustering and projection Robust to handle continuous, discrete values, sparse data etc.

Questions?