Subspace Clustering. Weiwei Feng. December 11, 2015

Similar documents
Generalized Principal Component Analysis CVPR 2007

Generalized Transitive Distance with Minimum Spanning Random Forest

Dimension Reduction CS534

Low Rank Representation Theories, Algorithms, Applications. 林宙辰 北京大学 May 3, 2012

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING

On Order-Constrained Transitive Distance

A Geometric Analysis of Subspace Clustering with Outliers

A new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity

Image Processing. Image Features

Principal Coordinate Clustering

Manifold Clustering. Abstract. 1. Introduction

Subspace Clustering for Sequential Data

Dimension Reduction of Image Manifolds

Learning a Manifold as an Atlas Supplementary Material

Recognition: Face Recognition. Linda Shapiro EE/CSE 576

Large-Scale Face Manifold Learning

Application of Spectral Clustering Algorithm

Non-linear dimension reduction

Data fusion and multi-cue data matching using diffusion maps

Machine Learning : Clustering, Self-Organizing Maps

Technical Report. Title: Manifold learning and Random Projections for multi-view object recognition

CSE 6242 A / CX 4242 DVA. March 6, Dimension Reduction. Guest Lecturer: Jaegul Choo

Robust Pose Estimation using the SwissRanger SR-3000 Camera

Computational Statistics and Mathematics for Cyber Security

Network Traffic Measurements and Analysis

LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA

Modelling and Visualization of High Dimensional Data. Sample Examination Paper

Time Series Clustering Ensemble Algorithm Based on Locality Preserving Projection

Latent Space Sparse Subspace Clustering

MSA220 - Statistical Learning for Big Data

School of Computer and Communication, Lanzhou University of Technology, Gansu, Lanzhou,730050,P.R. China

Discriminate Analysis

Graph Connectivity in Sparse Subspace Clustering

Linear Discriminant Analysis for 3D Face Recognition System

Robust Subspace Clustering via Half-Quadratic Minimization

The Pre-Image Problem in Kernel Methods

CPSC 340: Machine Learning and Data Mining. Multi-Dimensional Scaling Fall 2017

Data Mining Chapter 3: Visualizing and Exploring Data Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

3D Motion Segmentation Using Intensity Trajectory

Robust Trajectory Clustering for Motion Segmentation

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS

A Course in Machine Learning

Latent Low-Rank Representation

Subspace Clustering with Global Dimension Minimization And Application to Motion Segmentation

Chapter 7. Conclusions and Future Work

Face Recognition for Mobile Devices

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

A Taxonomy of Semi-Supervised Learning Algorithms

Mathematics of Data. INFO-4604, Applied Machine Learning University of Colorado Boulder. September 5, 2017 Prof. Michael Paul

Intro to Artificial Intelligence

Emotion Classification

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Three-Dimensional Motion Tracking using Clustering

CS664 Lecture #18: Motion

Neural Networks for Machine Learning. Lecture 15a From Principal Components Analysis to Autoencoders

Subspace Clustering by Block Diagonal Representation

Maths for Signals and Systems Linear Algebra in Engineering. Some problems by Gilbert Strang

Facial Expression Recognition using Principal Component Analysis with Singular Value Decomposition

Modelling Data Segmentation for Image Retrieval Systems

Announcements. Recognition I. Gradient Space (p,q) What is the reflectance map?

Spectral Surface Reconstruction from Noisy Point Clouds

CSE 6242 / CX October 9, Dimension Reduction. Guest Lecturer: Jaegul Choo

A Generalized Method to Solve Text-Based CAPTCHAs

10-701/15-781, Fall 2006, Final

Locally Linear Landmarks for large-scale manifold learning

Factorization with Missing and Noisy Data

Diffusion Wavelets for Natural Image Analysis

Sparse Manifold Clustering and Embedding

Clustering: Classic Methods and Modern Views

Selecting Models from Videos for Appearance-Based Face Recognition

Accelerometer Gesture Recognition

Subspace Clustering by Block Diagonal Representation

Unsupervised learning in Vision

CS 229 Midterm Review

Basis Functions. Volker Tresp Summer 2017

Fuzzy C-means Clustering with Temporal-based Membership Function

Learning based face hallucination techniques: A survey

Localized K-flats. National University of Defense Technology, Changsha , China wuyi

calibrated coordinates Linear transformation pixel coordinates

An Intelligent Clustering Algorithm for High Dimensional and Highly Overlapped Photo-Thermal Infrared Imaging Data

REPRESENTATION OF BIG DATA BY DIMENSION REDUCTION

arxiv: v1 [cs.cv] 18 Jan 2015

Topological Data Analysis

CS 534: Computer Vision Segmentation II Graph Cuts and Image Segmentation

Segmentation of MR Images of a Beating Heart

Head Frontal-View Identification Using Extended LLE

Learning from High Dimensional fmri Data using Random Projections

Statistics 202: Data Mining. c Jonathan Taylor. Week 8 Based in part on slides from textbook, slides of Susan Holmes. December 2, / 1

Scalable Sparse Subspace Clustering

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

arxiv: v1 [cs.cv] 9 Sep 2013

Final Review CMSC 733 Fall 2014

Feature selection. Term 2011/2012 LSI - FIB. Javier Béjar cbea (LSI - FIB) Feature selection Term 2011/ / 22

Facial Expression Detection Using Implemented (PCA) Algorithm

Content-based Image and Video Retrieval. Image Segmentation

Geodesic Flow Kernel for Unsupervised Domain Adaptation

Summer School: Mathematical Methods in Robotics

Chap.12 Kernel methods [Book, Chap.7]

Transcription:

Subspace Clustering Weiwei Feng December 11, 2015 Abstract Data structure analysis is an important basis of machine learning and data science, which is now widely used in computational visualization problems, e.g. facial recognition, image classification, and motion segmentation. In this project, I would like to deal with a set of small classification problems and use methods like PCA, spectral analysis, kmanifold, etc. By exploring different methods, I would try to prove both mathematically and in application that different problems need to be treated differently, and use their case specific methods. Keywords. independent subspace, nonlinear subspace, PCA, clustering analysis, Kmanifold algorithm, KNN,SSC,LSC 1 problem background and analysis A successful analysis on manifold data analysis is based on appropriately dealt with its data structure, and design of specific algorithm. As one of the main topics on manifold learning, subspace clustering has bright prospect in application. According to different assumptions, it is classified as Sparse Subspace Clustering(SSC) and Low-rank Subspace Clustering (LSC). First of all, the two methods both assume that the observed data satisfy X = X 0 + E R d n, there are n d dimensional data points. X 0 is clean (non-noised) data, E is noise or extreme value indispensable in the data collection process. The first assumption of subspace clustering is to assume that data points can be expressed using a few vector bases. Mathematically, we can write the assumption as X 0 = AZ, here A is a matrix, Z is sparse matrix. In the following algorithm, we use A = X O, which implies that the data is auto-representative. The difference between two methods is that SSC assumes the matrix Z is sparse, i.e. there are only a few effective representative vectors. While LSC assumes that the matrix Z is low-rank. So for SSC, we have minimize 1 2 X XZ 2 F + λ Z 1 subjet to diag(z) = 0. and for LSC, we have minimize Z + λ X XZ 1. 1

The Kmanifold algorithm, which I used a lot in this project, is based on Isomap and EM algorithm. Here is the realization of Kmanifold algorithm. Algorithm 1 Kmanifold Require: A set of images {X 1, X 2,..., X n } and a desired label set size, k 1: Calculate the Euclidean distance between any of the two points in the data set 2: Create a graph G with a node for each image, and an edge between pairs of neighboring images and set the edge weight to that distance 3: Compute all pairs shortest path distances on G. Let d(i, j) be the length of the shortest path between node i and j. 4: Initialization: Create a k n matrix W where w c i is the probability of X i belonging to manifold c. W can be inititalized randomly unless domain-specific priors are available. 5: M-Step: For each class, c, supply w c and D and use node-weighted MDS to embed the points in the desired dimension. 6: E-Step: For each point, estimate the distance to the manifolds implied by the output of Step 5 and re-weight accordingly. 7: Go to Step 5 until convergence. 2 problem solving Here I first tried to dig into the data structure of each problem, and then applied methods which are relatively robust, easy, and computationally feasible. 2.1 Problem I Data of Problem I are from two independent subspaces. From the geometric point of view, linear subspace are points, lines, planes,etc through the origin. Perpendicular linear subspaces imply that the two spaces are orthogonal. Hence, from the linear point of view, independent subspace are easy to differentiate. Common method like Principle Component Analysis (PCA) are suitable for solving this kind of problems, besides, PCA requires few computation. After obtaining the principle components, using simple clustering method, like K-means method, we are capable to separate data fro different subspaces. I did PCA on the data, and projected it onto the two dimensional space. See graph 1 (a). Obviously, data are well classified to two classes, and using K-means method, I got a result presented in graph 1(b). 2.2 Problem II For this problem, I used four different dataset which are all no longer simple independent linear classification problems. Hence here, PCA is not suitable. We analyze respectively these problems and then introduce proper algorithm to realize the clustering objective. 2.2.1 Problem II (a) The difficulty of this problem is that both of the two lines do not go through the origin, therefore it s not a subspace. Use subspace clustering method, will add one more affine, and reduce the 2

Figure 1: Problem I(1) Original data PCA analysis; (2) K-means method algorithm efficiency. But we notice that these two lines are both one dimensional manifold, and perpendicularly intercept. Manifold clustering?kmanifold) is proper for this kind of problem. The basic idea is to first construct the manifold measurement using Isomap algorithm, then assume the observed data are from low-rank dimensioanl manifold plus noise, construct the probability model, finally get the result using classic EM algorithm. Particularly, due to the perpendicularity, the manifold separation is very clear, so the result of Kmanifold is good. Figure 2: Problem II(a)(1) Original data; (2) Kmanifold method 2.2.2 Problem II (b) Data for this problem are from manifolds in three dimensional spaces. It looks harder then the previous problem, but all the spaces go through origin, so it perfectly satisfies the assumption of subspace clustering. We have found that lack of independency, SSC has more robust performance than LSC. So here I used SSC. I took regular parameter λ = 0.1, similarity matrix W = Z + Z. Results are as follows. 3

Figure 3: Problem II(b)(1) Original data; (2) SSC method 2.2.3 Problem II (c) We see that the two parabolas are nonlinear, one of the two not passes the origin, and the two do not intersect. This implies two points: 1. This problem doesn t satisfy the subspace clustering assumption. 2. The two one dimensional manifold do not intersect, so Kmanifold doesn t apply. Considering these two insights, we first use KNN to construct an affinity matrix, then use spectrum clustering method. I took K= 10. One thing to notice is that K cannot be too large, otherwise we violate the fact that the two lines do not intersect. Results are as follows. Figure 4: Problem II(c)(1) Original data; (2) LSC method 3 Conclusion From the previous parts, we can see that kmanifold, with the realization of SSC and LSC, can be applied to 2D and 3D data separation and classification. With smaller data set and simple structure, methods like PCA and KNN are both sufficient for solving our problems. However, with 4

bigger data set and previously unknown data structure, kmanifold is an efficient way of dealing with these data sets. Taking into consideration of the curse of dimensionality, LSC is more robust in multi-dimensional data problems. ( This interesting observation can be compared to QDA and LDA in their application with different data sets). 5

References [1] R. Basri and D. W. Jacobs. Lambertian reflectance and linear subspaces. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(2):218 233, 2003. [2] R. Vidal. Subspace clustering. IEEE Signal Processing Magazine, 28(2):52 68, 2011. [3] J. Shi and J. Malik, Normalized cuts and image segmentation. IEEE Transactions Pattern Analysis Machine Intelligence, 22(8):888 905, 2000. [4] G. Liu, Z. Lin, S. Yan, J. Sun, Y. Yu, and Y. Ma. Robust recovery of subspace structures by low-rank representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(1):171 184, 2013. [5] E. Elhamifar and R. Vidal. Sparse subspace clustering: Algorithm, theory, and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(11):2765 2781, 2013. 6