Learning Robust Low-Rank Representations
|
|
- Barbara Logan
- 6 years ago
- Views:
Transcription
1 Learning Robust Low-Rank Representations Pablo Sprechmann Dept. of Elec. and Comp. Eng. Duke University Alexander M. Bronstein School of Elec. Eng. Tel Aviv University Guillermo Sapiro Dept. of Elec. and Comp. Eng. and Dept. of Comp. Scien. Duke University Abstract In this paper we present a comprehensive framework for learning robust low-rank representations by combining and extending recent ideas for learning fast sparse coding regressors with structured non-convex optimization techniques. This approach connects robust principal component analysis (RPCA) with dictionary learning methods and allows its approximation via trainable encoders. We propose an efficient feed-forward architecture derived from an optimization algorithm designed to exactly solve robust low dimensional projections. This architecture, in combination with different training objective functions, allows the regressors to be used as online approximants of the exact offline RPCA problem or as RPCA-based neural networks. Simple modifications of these encoders can handle challenging extensions, such as the inclusion of geometric data transformations. We present several examples with real data from audio and video processing. When used to approximate RPCA, our basic implementation shows several orders of magnitude speedup compared to the exact solvers with almost no performance degradation. We show the strength of the inclusion of learning into the RPCA approach on a music source separation application, where the encoders outperform the exact RPCA algorithms, which are already reported to produce state-of-the-art results on a benchmark database. Video applications are demonstrated as well. Our preliary implementation on an ipad shows faster-than-real-time performance with imal latency. Introduction Principal component analysis (PCA) is the most widely used statistical technique for dimensionality reduction. However, its performance is highly sensitive to the presence of outliers, that is, samples not following the underlying low rank model. In a series of recent works Candès et al. (20) and Xu et al. (200) developed an elegant solution to this problem, in which the low rank matrix is detered as the imizer of a convex program. In this work, we concentrate in the noisy version of robust PCA (RPCA) obtained by solving the unconstrained program L,O R m n 2 X L O 2 F + λ L + λ O, () where X R m n is the data matrix, L is a low rank matrix and O is an error matrix with a sparse number of non-zero coefficients (with arbitrarily large magnitude). L denotes the matrix nuclear norm, defined as the sum of the singular values of L. The parameters λ and λ are a positive scalar parameters controlling the rank of L and the sparsity level in the outliers, respectively, and can be set automatically as in Ramirez & Sapiro (202). This particular formulation of RPCA has attracted significant interest in the machine learning, computer vision, and signal processing communities, and was successfully used in various applications such as Wagner et al. (20); Peng et al. (200); Zhou et al. (200); Mu et al. (20). A challenge often encountered in modern settings is that the flow of new input data is permanent. Then, the
2 robust low rank model needs to be adapted constantly since the principal directions can change over time, calling for developing efficient online techniques recently proposed in Qiu & Vaswani (20); Tan et al. (20); Mateos & Giannakis (202); Balzano et al. (200). Significant amount of effort has been devoted to developing optimization algorithms for efficiently solving () and its noiseless formulation Candès et al. (20); Lin et al. (2009); Ma et al. (20); Recht & Ré (20). Despite the permanent progress reported in the literature, state-of-the-art algorithms still have prohibitive complexity and latency for real-time processing. In the sparse coding domain, a similar situation was encountered a few years ago. This motivated significant effort in the deep learning community, e.g., by Ranzato et al. (2007) and Goodfellow et al. (2009), aig at overcog this problem. Works by Jarrett et al. (2009) and Kavukcuoglu et al. (200) proposed learning non-linear regressors capable of producing good approximations of the true sparse codes in a fixed amount of time. Gregor & LeCun (200) introduced an approach in which the regressors are multilayer artificial neural networks with an architecture inspired by first order optimization algorithms for solving sparse coding problems. This work was later extended for handling more general structured sparse models by Sprechmann et al. (202a). With this motivation, in this paper we propose to extend these ideas to the RPCA context. We propose to design regressors capable of approximating online RPCA in a very fast way. To the best of our knowledge, this type of encoders have never been developed before. Our RPCA encoders are learned by imizing various objective functions that allow their use in several different contexts. In Section 2, we present our approach to RPCA and discus exact optimization algorithms to solve it. In Section 3, we introduce the new robust encoders and the new objective functions used for their training. In Section 4, we present several experimental results. Conclusions are drawn in Section 5. For further details, the reader is referred to Sprechmann et al. (202b). 2 Online RPCA via non-convex factorization Using the result from Srebro & Shraibman (2005) and Recht et al. (200), Mateos & Giannakis (202) proposed to reformulate () as U,S,O 2 X US O 2 F + λ 2 ( U 2 F + S 2 F ) + λ O, (2) with U R m q, S R q n, and O R m n. Here, q is rough upper bound on rank(l) q. This decomposition reveals a lot of structure hidden in the problem. The low rank component can now be thought of as an under-complete dictionary U, with q atoms, multiplied by a matrix S containing in its columns the corresponding coefficients for each data vector in X. This interpretation brings our problem close to that of dictionary learning in the sparse modeling domain; see Mairal et al. (2009). While this new factorized formulation drastically reduces the number of optimization variables, problem (2) is no longer convex. Fortunately, Mardani et al. (20) showed that any stationary point of (2), {U, S, O}, satisfying X US O 2 λ is an optimal solution of (2). Thus, problem (2) can be solved using an alternating imization or block coordinate scheme, in which the cost function is imized with respect to each individual optimization variable while keeping the other ones fixed, without the risk of falling into a stationary point that may not be globally optimal. This will be exploited to design our fast encoders. Robust low dimensional projections: Let us assume that we have already learned a low dimensional model, U R m q, from some data X US + O R m n. Suppose that we are given a new input vector x R m drawn from the same distribution as X. Then x can be decomposed as x Us + o, where Us represents the low dimensional component, n is a small perturbation and o is a sparse outlier vector. This can be done by solving s R q,o R m 2 x Us o λ 2 s λ o. (3) Unlike dictionary learning problems, e.g. in Mairal et al. (2009), here the columns of the dictionary U are not constrained to have unit norm. The robust low dimensional projection (3) is a convex program that can be solved using several methods. We are interested in choosing an optimization algorithm that can be further used to define the architecture of trainable encoders for simultaneously estimating s and o. With this in d, we choose to use the block-coordinate imization scheme. 2
3 The solution of (3) is given by s = (U T U λ I) U T (x t o) and o = π λ (ˆx Us), when fixing o and s respectively. Here π λ is the scalar soft-thresholding operator with parameter λ R m, which applies a soft-threshold λ i to each component of the input vector. In our case, λ = λ. Note that we can write the estimates in a recursive manner as, o k+ = π λ (b k ), b k+ = b k + W(o k+ o k ), where k is the iteration number, and W = UHU T and H = (U T U λ I). After convergence, s can be obtained as s = H(x o). Online RPCA: So far, we assumed that the entire data matrix X was available a priori. We now address the case when the data samples {x t } t N, x t R m, arrive sequentially; the index t should be understood as a temporal variable. Online RPCA aims at estimating and refining the model as the data come in. An alternating imization algorithm for solving the online counterpart of (2) goes as follows: When a new data vector x t is received, we first obtain its representation {s t, o t } as a low dimensional projection (3) given the current model estimate, U t, which can be done using strategies such as block-coordinate descent methods with warm restarts described Mairal et al. (2009), or recursive least squares Mateos & Giannakis (202). 3 Online RPCA via fast trainable encoders The main contributions of this paper is the construction of trainable regressors that can be used for approximating the solution of (3). The main idea is to build a parametric regressor z = (s, o) = h(x, Θ), with some set of parameters, collectively denoted as Θ. Thus, we need to define an architecture for h and a learning algorithm in order to detere Θ. Following the fast sparse coding methods in Gregor & LeCun (200) and Sprechmann et al. (202a), we propose to use feed-forward multi-layer architecture where each layer implements a single iteration of the block-coordinate imization scheme described in Section 2. The parameters of the network are the matrices W and H and the thresholds λ (extra flexibility is obtained by learning different thresholds λ i for each component). The encoder architecture is depicted in Sprechmann et al. (202b). Each layer essentially consists of the nonlinear thresholding operator π λ followed by a linear operation W. The network parameters are initialized to replicate a fixed number of iterations of the exact algorithm. As a learning strategy, we propose to select the set of parameters Θ that imizes the loss function, L(Θ) = n L(Θ, x j ) (4) n on a training set X = {x,..., x n }. Here, L(Θ, x j ) is a function that measures the quality of the code z j = h(x j, Θ) produced for the data point x j. The selection of the objective function L sets the type of regressor that we are going to obtain and it is application-dependent. One of the simplest choices is to use L(Θ, x j ) = z j z j, with z j = (s j, o j ) being the j-th columns of the decomposition of the data X = (x,..., x n ) into X = US + O by the exact RPCA. This essentially trains the encoder to approximate the exact solution of the RPCA problem. In many applications, the data may not completely adhere to the assumptions of the RPCA model, and the exact solution is, therefore, not necessarily the best one. This is the case, for example, in the source separation problem presented in Huang et al. (202), where RPCA gives a very good separation of the spectrally sparse singing voice (modeled as outliers) and repetitive low-rank background accompaniment, yet the obtained signals are still not equal to the true voice and background tracks. In this case, one could use a collection of clean voice and background tracks, {s j } and {o j } respectively, to supervise the training; this leads to better separation results as reported in Section 4. In a fully online setting, one does not have access either to the exact RPCA decomposition nor the true ground truth separation. In that case, we propose to directly use L(Θ, x j ) = f(x, z), where f is the objective function in the low dimensional projection (3). In this fully unsupervised case the encoders can be viewed as stand alone RPCA regressors, which brings them close in spirit to the sparse auto-encoders proposed by Goodfellow et al. (2009). Several applications of RPCA rely on the critical assumption that the given input vectors are aligned with respect to a group of geometric transformations. Translating the work by Peng Same could be done solving (3) using proximal methods instead of block-coordinate imization. j= 3
4 Figure : Robust PCA representation of several frames (top-to-bottom) from the surveillance sequence obtained using the algorithm from Lin et al. (2009) (left), and a five layer network encoder (right). Columns in each group are, leftto-right: the reconstructed frame, its lowrank approximation (background), and the sparse outlier (foreground). Method GNSDR GSNR GSAR GSIR Ideal freq. mask ADMoM RPCA Huang et al. (202) Proximal RPCA NN RPCA Untrained NN RPCA Unsupervised NN RPCA Supervised Table : Performance of audio separation methods on the MIR-K dataset. et al. (200) to our setting, we propose to use the training objective function defined in (4) with L Tr (Θ, x j, α j ) = f(t αj (x j ), h(t αj (x j ), Θ)), where T α is a parametrized operator (with a set of parameters collectively denoted as α) that applies different geometric transformation, T αj, to each training vector x j. In the online setting, when a new data vector x t arrives, we compute its robust low rank projection by imizing L Tr (Θ, x t, α t ) over a the vector α t parameterizing the geometric transformation. In the same way, as new data arrive, the transformation of all the previously seen training vectors is updated through the imization of a loss function L(Θ, α) with respect to α. The imization of a loss function L(Θ) with respect to Θ or α is performed using a stochastic gradient descent, see Gregor & LeCun (200) and Sprechmann et al. (202a) for details. 4 Experimental results Video separation: Figure shows background and foreground separation via robust PCA on the surveillance video sequence Hall of a business building taken from Li et al. (2004). The sequence consists of images of an indoor scene shot by a static camera in a mall. The scene has a nearly constant background and walking people in the foreground. We used networks with five layers and q = 5 trained to approximate the output of the exact RPCA on a subset of the frames in the sequence. The separation produced by the fast encoder is nearly identical to the output of the exact algorithm and to the output of the code from Lin et al. (2009), used as reference, while being considerably faster. Our Matlab implementation with built-in GPU acceleration executed on an NVIDIA Tesla C2070 GPU propagates a frame through a single layer of the network in merely 92µsec. This is several orders of magnitude faster than iterative solver executed on the CPU. Audio separation: We evaluate the separation performance of the proposed methods on the MIR- K dataset from Hsu & Jang (200), containing 000 clips extraced from 0 Chinese karaoke songs. The experimental settings closely followed that of Lin et al. (2009), to which the reader is referred for further details. As the evaluation criteria, we used the BSS-EVAL metrics developed in Vincent et al. (2006), which calculate the global normalized source-to-distortion ratio (GNSDR),source-to-artifacts ratio (GSAR), source-to-interference ratio (GSIR), and signal-tonoise ratio (GSNR). All networks used 20 layers with q = 25. The following training regimes were compared: untrained parameters initialized according to Section 2 (Untrained); unsupervised learning (Unsupervised); and training supervised by the clean voice and background tracks (Supervised). Table summarizes the obtained separation performance. While unsupervised training makes fast RPCA encoders on par with the exact RPCA (at a fraction of the computational complexity and latency of the latter), significant improvement is achieved by using the supervised regime. 5 Conclusion By combining ideas from structured non-convex optimization with multi-layer neural networks, we have developed a comprehensive framework for the online learning of robust low-rank representations in real time and capable of handling large scale applications. A basic implementation already achieves several order of magnitude speedups when compared to exact solvers, opening the door for practical algorithms in various applications. 4
5 References Balzano, L., Nowak, R., and Recht, B. Online identification and tracking of subspaces from highly incomplete information. In Proc. of 48th Allerton Conf., 200. Candès, E., Li, X., Ma, Y., and Wright, J. Robust principal component analysis? Journal of the ACM, 58(3), 20. Goodfellow, I., Le, Q., Saxe, A., Lee, H., and Ng, A. Y. Measuring invariances in deep networks. In In NIPS, pp Gregor, K. and LeCun, Y. Learning fast approximations of sparse coding. In ICML, pp , 200. Hsu, C.L. and Jang, J.S.R. On the improvement of singing voice separation for monaural recordings using the MIR-K dataset. IEEE Trans. on Audio, Speech, and Lang. Proc., 8(2):30 39, 200. Huang, P-S., Chen, S.D., Smaragdis, P., and Hasegawa-Johnson, M. Singing-voice separation from monaural recordings using robust principal component analysis. In ICASSP, 202. Jarrett, K., Kavukcuoglu, K., Ranzato, M.A., and LeCun, Y. What is the best multi-stage architecture for object recognition? In CVPR, pp , Kavukcuoglu, K., Ranzato, M.A., and LeCun, Y. Fast inference in sparse coding algorithms with applications to object recognition. arxiv: , 200. Li, L., Huang, W., Gu, I. Yu-Hua, and Tian, Q. Statistical modeling of complex backgrounds for foreground object detection. IEEE Trans. Image Process., 3(): , Lin, Z., Ganesh, A., Wright, J., Wu, L., Chen, M., and Ma, Y. Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix. preprint, Ma, S., Goldfarb, D., and Chen, L. Fixed point and Bregman iterative methods for matrix rank imization. Math. Program., 28(-2):32 353, 20. Mairal, J., Bach, F., Ponce, J., and Sapiro, G. Online dictionary learning for sparse coding. In ICML, pp , Mardani, M., Mateos, G., and Giannakis, G. B. Unveiling network anomalies in large-scale networks via sparsity and low rank. In Proc. of 44th Asilomar Conf. on Signals, Systems, and Computers, 20. Mateos, G. and Giannakis, G. B. Robust PCA as bilinear decomposition with outlier-sparsity regularization. IEEE Trans. on Signal Process., 60(0): , 202. Mu, Y., Dong, J., Yuan, X., and Yan, S. Accelerated low-rank visual recovery by random projection. In CVPR, pp , 20. Peng, Y., Ganesh, A., Wright, J., Xu, W., and Ma, Y. RASL: Robust alignment by sparse and low-rank decomposition for linearly correlated images. In CVPR, pp , 200. Qiu, C. and Vaswani, N. Real-time robust principal components pursuit. arxiv.org:.788, 20. Ramirez, I. and Sapiro, G. Low-rank data modeling via the imum description length principle. In ICASSP, 202. Ranzato, M., Huang, F. J., Boureau, Y-L., and LeCun, Y. Unsupervised learning of invariant feature hierarchies with applications to object recognition. In CVPR, Recht, B. and Ré, C. Parallel stochastic gradient algorithms for large-scale matrix completion. 20. Recht, B., Fazel, M., and Parrilo, P. A. Guaranteed imum-rank solutions of linear matrix equations via nuclear norm imization. SIAM Rev., 52(3):47 50, 200. Sprechmann, P., Bronstein, A. M., and Sapiro, G. Learning efficient structured sparse models. In ICML, 202a. Sprechmann, P., Bronstein, A. M., and Sapiro, G. Learning Robust Low-Rank Representations. arxiv.org: , 202b. Srebro, N. and Shraibman, A. Rank, trace-norm and max-norm. In COLT, pp , Tan, W., Cheung, G., and Ma, Y. Face recovery in conference video streag using robust principal component analysis. In ICIP, pp , 20. Vincent, E., Gribonval, R., and Févotte, C. Performance measurement in blind audio source separation. IEEE Trans. on Audio, Speech, and Lang. Proc., 4(4): , Wagner, A., Wright, J., Ganesh, A., Zhou, Z., Mobahi, H., and Ma, Y. Towards a practical face recognition system: Robust alignment and illuation via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell., 34(2): , 20. Xu, H., Caramanis, C., and Sanghavi, S. Robust PCA via outlier pursuit. In NIPS, pp Zhou, Z., Li, X., Wright, J., Candès, E. J., and Ma, Y. Stable principal component pursuit. In ISIT, pp ,
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference
Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400
More informationRobust Principal Component Analysis (RPCA)
Robust Principal Component Analysis (RPCA) & Matrix decomposition: into low-rank and sparse components Zhenfang Hu 2010.4.1 reference [1] Chandrasekharan, V., Sanghavi, S., Parillo, P., Wilsky, A.: Ranksparsity
More informationSupplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision
Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary
More informationIMA Preprint Series # 2281
DICTIONARY LEARNING AND SPARSE CODING FOR UNSUPERVISED CLUSTERING By Pablo Sprechmann and Guillermo Sapiro IMA Preprint Series # 2281 ( September 2009 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS UNIVERSITY
More informationONLINE ROBUST PRINCIPAL COMPONENT ANALYSIS FOR BACKGROUND SUBTRACTION: A SYSTEM EVALUATION ON TOYOTA CAR DATA XINGQIAN XU THESIS
ONLINE ROBUST PRINCIPAL COMPONENT ANALYSIS FOR BACKGROUND SUBTRACTION: A SYSTEM EVALUATION ON TOYOTA CAR DATA BY XINGQIAN XU THESIS Submitted in partial fulfillment of the requirements for the degree of
More informationLimitations of Matrix Completion via Trace Norm Minimization
Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing
More informationFAST PRINCIPAL COMPONENT PURSUIT VIA ALTERNATING MINIMIZATION
FAST PRICIPAL COMPOET PURSUIT VIA ALTERATIG MIIMIZATIO Paul Rodríguez Department of Electrical Engineering Pontificia Universidad Católica del Perú Lima, Peru Brendt Wohlberg T-5 Applied Mathematics and
More informationUnsupervised learning in Vision
Chapter 7 Unsupervised learning in Vision The fields of Computer Vision and Machine Learning complement each other in a very natural way: the aim of the former is to extract useful information from visual
More informationCOMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS. Jingyu Yang, Xuemeng Yang, Xinchen Ye
COMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS Jingyu Yang, Xuemeng Yang, Xinchen Ye School of Electronic Information Engineering, Tianjin University Building
More informationOnline PLCA for Real-time Semi-supervised Source Separation
Online PLCA for Real-time Semi-supervised Source Separation Zhiyao Duan 1, Gautham J. Mysore 2 and Paris Smaragdis 2,3 1 EECS Department, Northwestern University, 2 Advanced Technology Labs, Adobe Systems
More informationLow Rank Representation Theories, Algorithms, Applications. 林宙辰 北京大学 May 3, 2012
Low Rank Representation Theories, Algorithms, Applications 林宙辰 北京大学 May 3, 2012 Outline Low Rank Representation Some Theoretical Analysis Solution by LADM Applications Generalizations Conclusions Sparse
More informationLearning based face hallucination techniques: A survey
Vol. 3 (2014-15) pp. 37-45. : A survey Premitha Premnath K Department of Computer Science & Engineering Vidya Academy of Science & Technology Thrissur - 680501, Kerala, India (email: premithakpnath@gmail.com)
More informationReal-time Background Subtraction via L1 Norm Tensor Decomposition
Real-time Background Subtraction via L1 Norm Tensor Decomposition Taehyeon Kim and Yoonsik Choe Yonsei University, Seoul, Korea E-mail: pyomu@yonsei.ac.kr Tel/Fax: +82-10-2702-7671 Yonsei University, Seoul,
More informationIN some applications, we need to decompose a given matrix
1 Recovery of Sparse and Low Rank Components of Matrices Using Iterative Method with Adaptive Thresholding Nematollah Zarmehi, Student Member, IEEE and Farokh Marvasti, Senior Member, IEEE arxiv:1703.03722v2
More informationarxiv: v1 [cs.cv] 9 Sep 2013
Learning Transformations for Clustering and Classification arxiv:139.274v1 [cs.cv] 9 Sep 213 Qiang Qiu Department of Electrical and Computer Engineering Duke University Durham, NC 2778, USA Guillermo Sapiro
More informationELEG Compressive Sensing and Sparse Signal Representations
ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /
More informationDirect Matrix Factorization and Alignment Refinement: Application to Defect Detection
Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection Zhen Qin (University of California, Riverside) Peter van Beek & Xu Chen (SHARP Labs of America, Camas, WA) 2015/8/30
More informationSection 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1
Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients
More informationBlockwise Matrix Completion for Image Colorization
Blockwise Matrix Completion for Image Colorization Paper #731 Abstract Image colorization is a process of recovering the whole color from a monochrome image given that a portion of labeled color pixels.
More informationReal-Time Principal Component Pursuit
Real-Time Principal Component Pursuit Graeme Pope, Manuel Baumann, Christoph Studer, and Giuseppe Durisi Dept. of Information Technology and Electrical Engineering, ETH Zurich, Zurich, Switzerland e-mails:
More informationNovel Lossy Compression Algorithms with Stacked Autoencoders
Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is
More informationSupplementary material for the paper Are Sparse Representations Really Relevant for Image Classification?
Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification? Roberto Rigamonti, Matthew A. Brown, Vincent Lepetit CVLab, EPFL Lausanne, Switzerland firstname.lastname@epfl.ch
More informationOutlier Pursuit: Robust PCA and Collaborative Filtering
Outlier Pursuit: Robust PCA and Collaborative Filtering Huan Xu Dept. of Mechanical Engineering & Dept. of Mathematics National University of Singapore Joint w/ Constantine Caramanis, Yudong Chen, Sujay
More informationAn Approach for Reduction of Rain Streaks from a Single Image
An Approach for Reduction of Rain Streaks from a Single Image Vijayakumar Majjagi 1, Netravati U M 2 1 4 th Semester, M. Tech, Digital Electronics, Department of Electronics and Communication G M Institute
More informationRobust Transfer Principal Component Analysis with Rank Constraints
Robust Transfer Principal Component Analysis with Rank Constraints Yuhong Guo Department of Computer and Information Sciences Temple University, Philadelphia, PA 19122, USA yuhong@temple.edu Abstract Principal
More informationarxiv: v1 [cs.cv] 1 Aug 2013
Domain-invariant Face Recognition using Learned Low-rank Transformation arxiv:1308.0275v1 [cs.cv] 1 Aug 2013 Abstract Qiang Qiu Duke University Durham, NC, 27708 qiang.qiu@duke.edu Guillermo Sapiro Duke
More informationLearning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro
Learning Low-rank Transformations: Algorithms and Applications Qiang Qiu Guillermo Sapiro Motivation Outline Low-rank transform - algorithms and theories Applications Subspace clustering Classification
More informationOnline Subspace Estimation and Tracking from Missing or Corrupted Data
Online Subspace Estimation and Tracking from Missing or Corrupted Data Laura Balzano www.ece.wisc.edu/~sunbeam Work with Benjamin Recht and Robert Nowak Subspace Representations Capture Dependencies Subspace
More informationFace Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN
2016 International Conference on Artificial Intelligence: Techniques and Applications (AITA 2016) ISBN: 978-1-60595-389-2 Face Recognition Using Vector Quantization Histogram and Support Vector Machine
More informationELEG Compressive Sensing and Sparse Signal Representations
ELEG 867 - Compressive Sensing and Sparse Signal Representations Introduction to Matrix Completion and Robust PCA Gonzalo Garateguy Depart. of Electrical and Computer Engineering University of Delaware
More informationPhotometric Stereo with Auto-Radiometric Calibration
Photometric Stereo with Auto-Radiometric Calibration Wiennat Mongkulmann Takahiro Okabe Yoichi Sato Institute of Industrial Science, The University of Tokyo {wiennat,takahiro,ysato} @iis.u-tokyo.ac.jp
More informationSparse Coding for Learning Interpretable Spatio-Temporal Primitives
Sparse Coding for Learning Interpretable Spatio-Temporal Primitives Taehwan Kim TTI Chicago taehwan@ttic.edu Gregory Shakhnarovich TTI Chicago gregory@ttic.edu Raquel Urtasun TTI Chicago rurtasun@ttic.edu
More informationSparse Coding and Dictionary Learning for Image Analysis
Sparse Coding and Dictionary Learning for Image Analysis Part IV: Recent Advances in Computer Vision and New Models Francis Bach, Julien Mairal, Jean Ponce and Guillermo Sapiro CVPR 10 tutorial, San Francisco,
More informationLEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA
LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES Qiang Qiu and Guillermo Sapiro Duke University, Durham, NC 2778, USA ABSTRACT Learning a transformation-based dimension reduction, thereby compressive,
More informationarxiv: v1 [eess.sp] 1 Nov 2017
LEARNED CONVOLUTIONAL SPARSE CODING Hillel Sreter Raja Giryes School of Electrical Engineering, Tel Aviv University, Tel Aviv, Israel, {hillelsr@mail., raja@}tau.ac.il arxiv:7.00328v [eess.sp] Nov 207
More informationSparse Shape Registration for Occluded Facial Feature Localization
Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature
More informationBilevel Sparse Coding
Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional
More informationPatch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques
Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques Syed Gilani Pasha Assistant Professor, Dept. of ECE, School of Engineering, Central University of Karnataka, Gulbarga,
More informationIMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING
IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING Jianzhou Feng Li Song Xiaog Huo Xiaokang Yang Wenjun Zhang Shanghai Digital Media Processing Transmission Key Lab, Shanghai Jiaotong University
More informationMULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo
MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS Yanghao Li, Jiaying Liu, Wenhan Yang, Zongg Guo Institute of Computer Science and Technology, Peking University, Beijing, P.R.China,
More informationFault detection with principal component pursuit method
Journal of Physics: Conference Series PAPER OPEN ACCESS Fault detection with principal component pursuit method Recent citations - Robust Principal Component Pursuit for Fault Detection in a Blast Furnace
More informationEffectiveness of Sparse Features: An Application of Sparse PCA
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationConvex Optimization / Homework 2, due Oct 3
Convex Optimization 0-725/36-725 Homework 2, due Oct 3 Instructions: You must complete Problems 3 and either Problem 4 or Problem 5 (your choice between the two) When you submit the homework, upload a
More informationADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING
ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING Fei Yang 1 Hong Jiang 2 Zuowei Shen 3 Wei Deng 4 Dimitris Metaxas 1 1 Rutgers University 2 Bell Labs 3 National University
More informationCOMP 551 Applied Machine Learning Lecture 16: Deep Learning
COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all
More informationWeighted-CS for reconstruction of highly under-sampled dynamic MRI sequences
Weighted- for reconstruction of highly under-sampled dynamic MRI sequences Dornoosh Zonoobi and Ashraf A. Kassim Dept. Electrical and Computer Engineering National University of Singapore, Singapore E-mail:
More informationCPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016
CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:
More informationA Geometric Analysis of Subspace Clustering with Outliers
A Geometric Analysis of Subspace Clustering with Outliers Mahdi Soltanolkotabi and Emmanuel Candés Stanford University Fundamental Tool in Data Mining : PCA Fundamental Tool in Data Mining : PCA Subspace
More informationSparsity and image processing
Sparsity and image processing Aurélie Boisbunon INRIA-SAM, AYIN March 6, Why sparsity? Main advantages Dimensionality reduction Fast computation Better interpretability Image processing pattern recognition
More informationSPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari
SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari Laboratory for Advanced Brain Signal Processing Laboratory for Mathematical
More informationNTHU Rain Removal Project
People NTHU Rain Removal Project Networked Video Lab, National Tsing Hua University, Hsinchu, Taiwan Li-Wei Kang, Institute of Information Science, Academia Sinica, Taipei, Taiwan Chia-Wen Lin *, Department
More informationADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.
ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N. Dartmouth, MA USA Abstract: The significant progress in ultrasonic NDE systems has now
More informationModeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations
Modeling Visual Cortex V4 in Naturalistic Conditions with Invariant and Sparse Image Representations Bin Yu Departments of Statistics and EECS University of California at Berkeley Rutgers University, May
More informationFace Recognition At-a-Distance Based on Sparse-Stereo Reconstruction
Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,
More informationImage Restoration and Background Separation Using Sparse Representation Framework
Image Restoration and Background Separation Using Sparse Representation Framework Liu, Shikun Abstract In this paper, we introduce patch-based PCA denoising and k-svd dictionary learning method for the
More informationProceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong
, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA
More informationSupervised Translation-Invariant Sparse Coding
Supervised Translation-Invariant Sparse Coding Jianchao Yang,KaiYu, Thomas Huang Beckman Institute, University of Illinois at Urbana-Champaign NEC Laboratories America, Inc., Cupertino, California {jyang29,
More informationConsistent dictionary learning for signal declipping
Consistent dictionary learning for signal declipping Lucas Rencker 1 ( ), Francis Bach, Wenwu Wang 1, and Mark D. Plumbley 1 1 Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford,
More informationOR-PCA with MRF for Robust Foreground Detection in Highly Dynamic Backgrounds
OR-PCA with MRF for Robust Foreground Detection in Highly Dynamic Backgrounds Sajid Javed 1, Seon Ho Oh 1, Andrews Sobral 2, Thierry Bouwmans 2 and Soon Ki Jung 1 1 School of Computer Science and Engineering,
More informationExtended Dictionary Learning : Convolutional and Multiple Feature Spaces
Extended Dictionary Learning : Convolutional and Multiple Feature Spaces Konstantina Fotiadou, Greg Tsagkatakis & Panagiotis Tsakalides kfot@ics.forth.gr, greg@ics.forth.gr, tsakalid@ics.forth.gr ICS-
More informationStacked Denoising Autoencoders for Face Pose Normalization
Stacked Denoising Autoencoders for Face Pose Normalization Yoonseop Kang 1, Kang-Tae Lee 2,JihyunEun 2, Sung Eun Park 2 and Seungjin Choi 1 1 Department of Computer Science and Engineering Pohang University
More informationRobust and Secure Iris Recognition
Robust and Secure Iris Recognition Vishal M. Patel University of Maryland, UMIACS pvishalm@umiacs.umd.edu IJCB 2011 Tutorial Sparse Representation and Low-Rank Representation for Biometrics Outline Iris
More informationChange-Point Estimation of High-Dimensional Streaming Data via Sketching
Change-Point Estimation of High-Dimensional Streaming Data via Sketching Yuejie Chi and Yihong Wu Electrical and Computer Engineering, The Ohio State University, Columbus, OH Electrical and Computer Engineering,
More informationCLEANING UP TOXIC WASTE: REMOVING NEFARIOUS CONTRIBUTIONS TO RECOMMENDATION SYSTEMS
CLEANING UP TOXIC WASTE: REMOVING NEFARIOUS CONTRIBUTIONS TO RECOMMENDATION SYSTEMS Adam Charles, Ali Ahmed, Aditya Joshi, Stephen Conover, Christopher Turnes, Mark Davenport Georgia Institute of Technology
More informationLearning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li
Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,
More informationConvex optimization algorithms for sparse and low-rank representations
Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,
More informationShow, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks
Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Zelun Luo Department of Computer Science Stanford University zelunluo@stanford.edu Te-Lin Wu Department of
More informationAdvanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude
Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,
More informationDiscriminative sparse model and dictionary learning for object category recognition
Discriative sparse model and dictionary learning for object category recognition Xiao Deng and Donghui Wang Institute of Artificial Intelligence, Zhejiang University Hangzhou, China, 31007 {yellowxiao,dhwang}@zju.edu.cn
More informationPartial Sum Minimization of Singular Values in RPCA for Low-Level Vision
2013 IEEE International Conference on Computer Vision Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Tae-Hyun Oh* KAIST Hyeongwoo Kim KAIST Yu-Wing Tai KAIST Jean-Charles Bazin
More informationSeismic data reconstruction with Generative Adversarial Networks
Seismic data reconstruction with Generative Adversarial Networks Ali Siahkoohi 1, Rajiv Kumar 1,2 and Felix J. Herrmann 2 1 Seismic Laboratory for Imaging and Modeling (SLIM), The University of British
More informationNULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING
NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING Pan Ji, Yiran Zhong, Hongdong Li, Mathieu Salzmann, Australian National University, Canberra NICTA, Canberra {pan.ji,hongdong.li}@anu.edu.au,mathieu.salzmann@nicta.com.au
More informationThe Benefit of Tree Sparsity in Accelerated MRI
The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients
More informationGeneralized Lasso based Approximation of Sparse Coding for Visual Recognition
Generalized Lasso based Approximation of Sparse Coding for Visual Recognition Nobuyuki Morioka The University of New South Wales & NICTA Sydney, Australia nmorioka@cse.unsw.edu.au Shin ichi Satoh National
More informationSparse Models in Image Understanding And Computer Vision
Sparse Models in Image Understanding And Computer Vision Jayaraman J. Thiagarajan Arizona State University Collaborators Prof. Andreas Spanias Karthikeyan Natesan Ramamurthy Sparsity Sparsity of a vector
More informationVisual object classification by sparse convolutional neural networks
Visual object classification by sparse convolutional neural networks Alexander Gepperth 1 1- Ruhr-Universität Bochum - Institute for Neural Dynamics Universitätsstraße 150, 44801 Bochum - Germany Abstract.
More informationSlides adapted from Marshall Tappen and Bryan Russell. Algorithms in Nature. Non-negative matrix factorization
Slides adapted from Marshall Tappen and Bryan Russell Algorithms in Nature Non-negative matrix factorization Dimensionality Reduction The curse of dimensionality: Too many features makes it difficult to
More informationSparsity Based Regularization
9.520: Statistical Learning Theory and Applications March 8th, 200 Sparsity Based Regularization Lecturer: Lorenzo Rosasco Scribe: Ioannis Gkioulekas Introduction In previous lectures, we saw how regularization
More informationAn efficient algorithm for sparse PCA
An efficient algorithm for sparse PCA Yunlong He Georgia Institute of Technology School of Mathematics heyunlong@gatech.edu Renato D.C. Monteiro Georgia Institute of Technology School of Industrial & System
More informationarxiv: v1 [cs.lg] 20 Dec 2013
Unsupervised Feature Learning by Deep Sparse Coding Yunlong He Koray Kavukcuoglu Yun Wang Arthur Szlam Yanjun Qi arxiv:1312.5783v1 [cs.lg] 20 Dec 2013 Abstract In this paper, we propose a new unsupervised
More informationAkarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction
Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different
More informationarxiv: v1 [cs.cv] 28 Sep 2018
Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,
More informationSupplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features
Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Sakrapee Paisitkriangkrai, Chunhua Shen, Anton van den Hengel The University of Adelaide,
More informationUniqueness in Bilinear Inverse Problems with Applications to Subspace and Sparsity Models
Uniqueness in Bilinear Inverse Problems with Applications to Subspace and Sparsity Models Yanjun Li Joint work with Kiryung Lee and Yoram Bresler Coordinated Science Laboratory Department of Electrical
More informationCombining Mask Estimates for Single Channel Audio Source Separation using Deep Neural Networks
Combining Mask Estimates for Single Channel Audio Source Separation using Deep Neural Networks Emad M. Grais, Gerard Roma, Andrew J.R. Simpson, Mark D. Plumbley Centre for Vision, Speech and Signal Processing,
More informationP257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies
P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies. azemi* (University of Alberta) & H.R. Siahkoohi (University of Tehran) SUMMARY Petrophysical reservoir properties,
More informationINTERACTIVE REFINEMENT OF SUPERVISED AND SEMI-SUPERVISED SOUND SOURCE SEPARATION ESTIMATES
INTERACTIVE REFINEMENT OF SUPERVISED AND SEMI-SUPERVISED SOUND SOURCE SEPARATION ESTIMATES Nicholas J. Bryan Gautham J. Mysore Center for Computer Research in Music and Acoustics, Stanford University Adobe
More informationIllumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model
Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model TAE IN SEOL*, SUN-TAE CHUNG*, SUNHO KI**, SEONGWON CHO**, YUN-KWANG HONG*** *School of Electronic Engineering
More informationSemi-supervised Data Representation via Affinity Graph Learning
1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073
More informationExploring Curve Fitting for Fingers in Egocentric Images
Exploring Curve Fitting for Fingers in Egocentric Images Akanksha Saran Robotics Institute, Carnegie Mellon University 16-811: Math Fundamentals for Robotics Final Project Report Email: asaran@andrew.cmu.edu
More informationBlind Identification of Graph Filters with Multiple Sparse Inputs
Blind Identification of Graph Filters with Multiple Sparse Inputs Santiago Segarra, Antonio G. Marques, Gonzalo Mateos & Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania
More informationModule 1 Lecture Notes 2. Optimization Problem and Model Formulation
Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization
More informationEE795: Computer Vision and Intelligent Systems
EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical
More informationSparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal
Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal Hadi. Zayyani, Seyyedmajid. Valliollahzadeh Sharif University of Technology zayyani000@yahoo.com, valliollahzadeh@yahoo.com
More informationDepartment of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2
Vol.3, Issue 3, 2015, Page.1115-1021 Effect of Anti-Forensics and Dic.TV Method for Reducing Artifact in JPEG Decompression 1 Deepthy Mohan, 2 Sreejith.H 1 PG Scholar, 2 Assistant Professor Department
More informationConditional gradient algorithms for machine learning
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationA low rank based seismic data interpolation via frequencypatches transform and low rank space projection
A low rank based seismic data interpolation via frequencypatches transform and low rank space projection Zhengsheng Yao, Mike Galbraith and Randy Kolesar Schlumberger Summary We propose a new algorithm
More informationBSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo
2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS Yongqin Zhang, Jiaying Liu, Mading Li, Zongming
More informationDeep Generative Models Variational Autoencoders
Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative
More informationOptimization. Industrial AI Lab.
Optimization Industrial AI Lab. Optimization An important tool in 1) Engineering problem solving and 2) Decision science People optimize Nature optimizes 2 Optimization People optimize (source: http://nautil.us/blog/to-save-drowning-people-ask-yourself-what-would-light-do)
More information