Low Rank Representation Theories, Algorithms, Applications. 林宙辰 北京大学 May 3, 2012

Similar documents
Robust Principal Component Analysis (RPCA)

Subspace Clustering. Weiwei Feng. December 11, 2015

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING

Latent Low-Rank Representation

Fixed-Rank Representation for Unsupervised Visual Learning

IN pattern analysis and signal processing, an underlying

arxiv: v1 [cs.cv] 18 Jan 2015

A REVIEW ON LOW-RANK MODELS IN DATA ANALYSIS. Zhouchen Lin. (Communicated by XXX)

Subspace Clustering by Block Diagonal Representation

COMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS. Jingyu Yang, Xuemeng Yang, Xinchen Ye

A Geometric Analysis of Subspace Clustering with Outliers

Subspace Clustering by Block Diagonal Representation

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Robust Subspace Segmentation with Block-diagonal Prior

Generalized Transitive Distance with Minimum Spanning Random Forest

Graph Connectivity in Sparse Subspace Clustering

Subspace Clustering for Sequential Data

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Robust Subspace Segmentation by Low-Rank Representation

Sparse Subspace Clustering for Incomplete Images

Robust Subspace Clustering via Half-Quadratic Minimization

Outlier Pursuit: Robust PCA and Collaborative Filtering

arxiv: v1 [cs.cv] 27 Apr 2014

Non-Negative Low Rank and Sparse Graph for Semi-Supervised Learning

Online Low-Rank Representation Learning for Joint Multi-subspace Recovery and Clustering

On Order-Constrained Transitive Distance

HIGH-DIMENSIONAL data are ubiquitous in many areas of

Learning a Manifold as an Atlas Supplementary Material

Latent Space Sparse Subspace Clustering

Complex Non-Rigid Motion 3D Reconstruction by Union of Subspaces

HIGH-dimensional data are commonly observed in various

LOW-RANK representation (LRR) [30], [33], as a promising

Oracle Based Active Set Algorithm for Scalable Elastic Net Subspace Clustering

ELEG Compressive Sensing and Sparse Signal Representations

A new Graph constructor for Semi-supervised Discriminant Analysis via Group Sparsity

Large-Scale Face Manifold Learning

Principal Coordinate Clustering

Unsupervised and Semi-Supervised Learning vial 1 -Norm Graph

IN some applications, we need to decompose a given matrix

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

Connections between the Lasso and Support Vector Machines

Analyzing Stochastic Gradient Descent for Some Non- Convex Problems

Pose Locality Constrained Representation for 3D Human Pose Reconstruction

Alternating Direction Method of Multipliers

Data-Dependent Sparsity for Subspace Clustering

Constructing a Non-Negative Low Rank and Sparse Graph with Data-Adaptive Features

Latent Space Sparse and Low-rank Subspace Clustering

ELEG Compressive Sensing and Sparse Signal Representations

FAST PRINCIPAL COMPONENT PURSUIT VIA ALTERNATING MINIMIZATION

arxiv: v1 [cs.cv] 8 Sep 2017

arxiv: v1 [cs.cv] 9 Sep 2013

Compressive Sensing. Part IV: Beyond Sparsity. Mark A. Davenport. Stanford University Department of Statistics

Real-time Background Subtraction via L1 Norm Tensor Decomposition

DescriptorEnsemble: An Unsupervised Approach to Image Matching and Alignment with Multiple Descriptors

Product Grassmann Manifold Representation and Its LRR Models

Generalized Principal Component Analysis via Lossy Coding and Compression Yi Ma

arxiv: v1 [cs.lg] 3 Jan 2018

Introduction to Topics in Machine Learning

arxiv: v2 [cs.cv] 7 Oct 2016

Fault detection with principal component pursuit method

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

Learning Robust Low-Rank Representations

Optic Cup Segmentation for Glaucoma Detection Using Low-Rank Superpixel Representation

Generalized Principal Component Analysis CVPR 2007

Graph Autoencoder-Based Unsupervised Feature Selection

Robust Subspace Segmentation by Simultaneously Learning Data Representations and Their Affinity Matrix

Distributed Low-rank Subspace Segmentation

Low-Rank Matrix Recovery III: Fast Algorithms and Scalable Applications Zhouchen Lin

Face Recognition via Sparse Representation

Robust l p -norm Singular Value Decomposition

arxiv: v1 [cs.cv] 13 Apr 2017

Random Walk Distributed Dual Averaging Method For Decentralized Consensus Optimization

IMA Preprint Series # 2281

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Shape Interaction Matrix Revisited and Robustified: Efficient Subspace Clustering with Corrupted and Incomplete Data

Learning based face hallucination techniques: A survey

FINDING a subset of a large number of models or data

Globally and Locally Consistent Unsupervised Projection

FINDING a subset of a large number of models or data points,

Batch image alignment via subspace recovery based on alternative sparsity pursuit

The Pre-Image Problem in Kernel Methods

Development in Object Detection. Junyuan Lin May 4th

Correntropy Induced L2 Graph for Robust Subspace Clustering

Signal Reconstruction from Sparse Representations: An Introdu. Sensing

60 2 Convex sets. {x a T x b} {x ã T x b}

Unsupervised Outlier Detection and Semi-Supervised Learning

arxiv: v1 [cs.cv] 12 Apr 2017

Bilevel Sparse Coding

Hashing with Graphs. Sanjiv Kumar (Google), and Shih Fu Chang (Columbia) June, 2011

Support vector machine (II): non-linear SVM. LING 572 Fei Xia

BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo

Lecture 17 Sparse Convex Optimization

A fast algorithm for sparse reconstruction based on shrinkage, subspace optimization and continuation [Wen,Yin,Goldfarb,Zhang 2009]

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation

Multiple-View Object Recognition in Band-Limited Distributed Camera Networks

Outline Introduction Problem Formulation Proposed Solution Applications Conclusion. Compressed Sensing. David L Donoho Presented by: Nitesh Shroff

A (somewhat) Unified Approach to Semisupervised and Unsupervised Learning

Data-Driven Geometry Processing Map Synchronization I. Qixing Huang Nov. 28 th 2018

Robust and Scalable Column/Row Sampling from Corrupted Big Data

Transcription:

Low Rank Representation Theories, Algorithms, Applications 林宙辰 北京大学 May 3, 2012

Outline Low Rank Representation Some Theoretical Analysis Solution by LADM Applications Generalizations Conclusions

Sparse Representation Sparse Representation m in jjxjj 0 ; s:t: y = A x: (1) Sparse Subspace Clustering m in jjz i jj 0 ; s:t: x i = X^iz i ; 8i: (2) where X^i = [x 1 ; ; x i 1 ; x i+1 ; ; x n ]. m in jjz jj 0 ; s:t: X = X Z ; diag(z ) = 0: (3) m in jjz jj 1 ; s:t: X = X Z ; diag(z ) = 0: E. Elhamifar and R. Vidal. Sparse Subspace Clustering. CVPR2009. (4)

Sparse Representation Construct a graph W = (jz j + j(z ) T j)=2 Normalized cut on the graph E. Elhamifar and R. Vidal. Sparse Subspace Clustering. CVPR2009.

Sparse Representation Theorem. Assum e the data is clean and is drawn from independent subspaces, then Z is block diagonal. dim ( P i S i ) = P i dim (S i ): E. Elhamifar and R. Vidal. Sparse Subspace Clustering. CVPR2009.

Drawback of SSC Sensitive to noise: no cross validation among coefficients m in jjz jj 1 ; s:t: X = X Z ; diag(z ) = 0: (4) m in jjz i jj 1 ; s:t: x i = X z i ; (z i ) i = 0: (5) E. Elhamifar and R. Vidal. Sparse Subspace Clustering. CVPR2009.

Hints from 2D Sparsity Rank is a good measure of 2D sparsity Real data usually lie on low-dim manifolds > 1B dim low-dim subspaces low rank data matrices Low rank high correlation among rows/columns

Low Rank Representation m in jjz jj 1 ; s:t: X = X Z ; diag(z ) = 0: (4) m in jjz jj ; s:t: X = X Z : no additional constraint! (6) jjz jj = P j ¾ j (Z ), nuclear norm, a convex surrogate of rank. Liu et al. Robust Subspace Segmentation by Low-Rank Representation, ICML 2010.

Low Rank Representation T heorem. A ssum e the data is clean and is draw n from indep endent subspaces, then there exists Z w hich is blo ck diagonal, and the rank of each blo ck equals the dim ension of the corresp onding subspace. Liu et al. Robust Subspace Segmentation by Low-Rank Representation, ICML 2010.

Low Rank Representation When there is noise and outliers m in jjz jj + jje jj 2;1 ; s:t: X = X Z + E : (7) where jje jj 2;1 = P i jje :;i jj 2. Liu et al. Robust Subspace Segmentation by Low-Rank Representation, ICML 2010.

Low Rank Representation Connection to Robust PCA m in ka k + jje jj 1 ; A ;E s:t: X = A + E ; (8) The clean data is low rank w.r.t. the dictionary I. m in jjz jj + jje jj 2;1 ; s:t: X = X Z + E : (7) The clean data is low rank w.r.t. the dictionary X. Candes, Li, Ma, and Wright, Robust Principal Component Analysis? Journal of the ACM, 2010.

Low Rank Representation Generalization m in jjz jj + jje jj 2;1 ; s:t: X = A Z + E : (9) The clean data is low rank w.r.t. the dictionary A. Liu et al. Robust Subspace Segmentation by Low-Rank Representation, ICML 2010.

More Theoretical Analysis Closed form solution at noiseless case m in kz k ; Z s:t: X = X Z ; has a unique closed-form optim al solution: Z = V r V T r skinny SV D of X. Shape Interaction Matrix, w here U r r V T r when X is sampled from independent subspaces, Z* is block diagonal, each block corresponding to a subspace is th e min kzk = rank(x): X=XZ S. Wei and Z. Lin. Analysis and Improvement of Low Rank Representation for Subspace segmentation, arxiv: 1107.1561, 2010. Liu et al. Robust Recovery of Subspace Structures by Low-Rank Representation, TPAMI 2012.

More Theoretical Analysis Closed form solution at general case m in Z kz k ; s:t: X = A Z ; has a unique closed-form optim al solution: Z = A y X. Liu et al., Robust Recovery of Subspace Structures by Low-Rank Representation, TPAMI 2012.

Follow-up Work More problems with closed form solutions min Z Speeding up optimization "kzk + 1 2 kxz Xk2 F min A;Z "kzk + 2 kaz Ak2 F + 1 2 kd Ak2 F min "kzk + 1 A;Z 2 ka Xk2 F ; s:t: A = AZ: P. Favaro, R. Vidal and A. Ravichandran, A Closed Form Solution to Robust Subspace Estimation and Clustering, CVPR 2011.

Exact Recoverability of LRR T heorem : L et = 3=(7kX k p n). T hen there exists > 0 such that w hen < =, LR R can exactly recover the row space and the colum n supp ort of (Z 0 ; E 0 ): U (U ) T = V 0 V T 0 ; I = I 0 ; w here = ji 0 j=n is the fraction of outliers, U is the colum n space of Z, V T 0 is the row space of Z 0, and I and I 0 is the colum n supp orts of E and E 0, resp ectively. Liu et al., Exact Subspace Segmentation and Outlier Detection by Low-Rank Representation, AISTATS 2012. Liu et al., Robust Recovery of Subspace Structures by Low-Rank Representation, TPAMI 2012.

Linearized Alternating Direction Model Problem ADM Difficulties min x;y Method (LADM) f(x) + g(y); s:t: A(x) + B(y) = c; where x, y and c could be either vectors or matrices, f and g are convex functions, and A and B are linear mappings. L A (x; y; ) = f(x) + g(y) + h ; A(x) + B(y) yi + 2 ka(x) + B(y) ck2 ; x k+1 = arg min x y k+1 = arg min y L A (x; y k ; k); L A (x k+1 ; y; k); k+1 = k + [A(x k+1 ) + B(y k+1 ) c]: Lin et al., Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation, NIPS 2011.

Linearized Alternating Direction Method (LADM) x k+1 = arg min f(x) + x 2 ka(x) + B(y k) c + k= k 2 ; y k+1 = arg min g(y) + y 2 ka(x k+1) + B(y) c + k= k 2 Linearize the quadratic term x k+1 = arg min x = arg min x y k+1 = arg min y Adaptive Penalty f(x) + ha ( k) + A (A(x k ) + B(y k ) c); x x k i + A 2 kx x k k 2 f(x) + A 2 kx x k + A ( k + (A(x k ) + B(y k ) c))=( A)k 2 ; g(y) + B 2 ky y k + B ( k + (A(x k+1 ) + B(y k ) c))=( B)k 2 :

Linearized Alternating Direction Method (LADM) Theorem: If f kg is non-decreasing and upper bounded, A > kak 2, and B > kbk 2, then the sequence f(x k ; y k ; k)g generated by LADMAP converges to a KKT point of the model problem. Lin et al., Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation, NIPS 2011.

Applying LADM to LRR LRR Further Acceleration Technique O(n 3 ) O(rn 2 ) min Z;E kzk + kek 2;1 ; s:t: X = XZ + E: Lin et al., Linearized Alternating Direction Method with Adaptive Penalty for Low-Rank Representation, NIPS 2011.

Experiments Table 1: Comparison among APG, ADM, LADM, standard LADMAP and accelerated LADMAP (denoted as LADMAP(A)) on the synthetic data. For each quadruple (s, p, d, ~r), the LRR problem, with ¹ = 0:1, was solved for the same data using di erent algorithms. We present typical running time (in 10 3 seconds), iteration number, relative error (%) of output solution ( ^E; ^Z) and the clustering accuracy (%) of tested algorithms, respectively. Size (s, p, d, ~r) Method Time Iter. (10, 20,200, 5) (15, 20,300, 5) (20, 25, 500, 5) (30, 30, 900, 5) k^z Z 0 k kz 0 k k ^E E 0 k ke 0 k Acc. APG 0.0332 110 2.2079 1.5096 81.5 ADM 0.0529 176 0.5491 0.5093 90.0 LADM 0.0603 194 0.5480 0.5024 90.0 LADMAP 0.0145 46 0.5480 0.5024 90.0 LADMAP(A) 0.0010 46 0.5480 0.5024 90.0 APG 0.0869 106 2.4824 1.0341 80.0 ADM 0.1526 185 0.6519 0.4078 83.7 LADM 0.2943 363 0.6518 0.4076 86.7 LADMAP 0.0336 41 0.6518 0.4076 86.7 LADMAP(A) 0.0015 41 0.6518 0.4076 86.7 APG 1.8837 117 2.8905 2.4017 72.4 ADM 3.7139 225 1.1191 1.0170 80.0 LADM 8.1574 508 0.6379 0.4268 80.0 LADMAP 0.7762 40 0.6379 0.4268 84.6 LADMAP(A) 0.0053 40 0.6379 0.4268 84.6 APG 6.1252 116 3.0667 0.9199 69.4 ADM 11.7185 220 0.6865 0.4866 76.0 LADM N.A. N.A. N.A. N.A. N.A. LADMAP 2.3891 44 0.6864 0.4294 80.1 LADMAP(A) 0.0058 44 0.6864 0.4294 80.1

Applications of LRR Image segmentation min Z;E kzk + kek 2;1 ; s:t: X = XZ + E: B. Cheng et al. Multi-task Low-rank Affinity Pursuit for Image Segmentation, ICCV 2011.

Saliency detection Applications of LRR m in kz k + ke k 2;1 ; Z ;E s:t: X = X Z + E : Lang et al. Saliency Detection by Multitask Sparsity Pursuit. IEEE TIP 2012.

Generalizations of LRR LRR with clean data min D ;Z;E kz k + ke k 2;1 ; s:t: D = D Z; X = D + E : m in D ;Z ;E kz k + ke k 2 F ; s:t: D = D Z ; X = D + E : S. Wei and Z. Lin. Analysis and Improvement of Low Rank Representation for Subspace segmentation, arxiv: 1107.1561, 2010. P. Favaro, R. Vidal and A. Ravichandran, A Closed Form Solution to Robust Subspace Estimation and Clustering, CVPR 2011.

Latent LRR Generalizations of LRR To address insufficient sample problem (X = [X ; X H ]Z O ;H =) X = X Z + LX ) m in Z ;L ;E kz k + klk + jje jj 2;1 ; s:t: X = X Z + LX + E : Liu and Yan. Latent Low-Rank Representation for Subspace Segmentation and Feature Extraction, ICCV 2011.

Generalizations of LRR Fixed Rank Representation (FRR) To address insufficient sample problem Subspace clustering: using m in Z ; ~ Z ;E Feature extraction: by m in Z ; ~ Z ;E ~Z kz ~ Z k 2 F + jje jj 2;1 ; s:t: X = X Z + E ; rank( ~ Z ) <= r: y = ~ Z x kz ~ Z k 2 F + jje jj 2;1 ; s:t: X = Z X + E ; rank( ~ Z ) <= r: Liu et al., Fixed-Rank Representation for Unsupervised Visual Learning, CVPR 2012.

Generalizations of LRR Semi-supervised learning min Z;E kz k + kz k 1 + ke k 2;1 ; s:t: X = AZ + E ; Z 0: propogate labels on the graph with weights (Z + (Z ) T )=2. L. Zhuang et al. Non-Negative Low Rank and Sparse Graph for Semi-Supervised Learning, CVPR 2012.

Kernel LRR Generalizations of LRR To address nonlinear multi-manifold segmentation m in Z ká(x ) Á(X )Z k 2 F + jjz jj : min Z X (z T i K (X ; X )z i 2K (x i ; X )z i + K (x i ; x i )) + jjz jj : i Kernel trick: há(x); Á(y)i = K (x; y) Wang et al., Structural Similarity and Distance in Learning, Annual Allerton Conf. Communication, Control and Computing 2011.

Thanks! zlin@pku.edu.cn m in kz k + ke k 2;1 ; Z ;E s:t: X = X Z + E : LRR = SSC + RPCA