Learning Robust Low-Rank Representations

Size: px
Start display at page:

Download "Learning Robust Low-Rank Representations"

Transcription

1 Learning Robust Low-Rank Representations Pablo Sprechmann Dept. of Elec. and Comp. Eng. Duke University Alexander M. Bronstein School of Elec. Eng. Tel Aviv University Guillermo Sapiro Dept. of Elec. and Comp. Eng. and Dept. of Comp. Scien. Duke University Abstract In this paper we present a comprehensive framework for learning robust low-rank representations by combining and extending recent ideas for learning fast sparse coding regressors with structured non-convex optimization techniques. This approach connects robust principal component analysis (RPCA) with dictionary learning methods and allows its approximation via trainable encoders. We propose an efficient feed-forward architecture derived from an optimization algorithm designed to exactly solve robust low dimensional projections. This architecture, in combination with different training objective functions, allows the regressors to be used as online approximants of the exact offline RPCA problem or as RPCA-based neural networks. Simple modifications of these encoders can handle challenging extensions, such as the inclusion of geometric data transformations. We present several examples with real data from audio and video processing. When used to approximate RPCA, our basic implementation shows several orders of magnitude speedup compared to the exact solvers with almost no performance degradation. We show the strength of the inclusion of learning into the RPCA approach on a music source separation application, where the encoders outperform the exact RPCA algorithms, which are already reported to produce state-of-the-art results on a benchmark database. Video applications are demonstrated as well. Our preliary implementation on an ipad shows faster-than-real-time performance with imal latency. Introduction Principal component analysis (PCA) is the most widely used statistical technique for dimensionality reduction. However, its performance is highly sensitive to the presence of outliers, that is, samples not following the underlying low rank model. In a series of recent works Candès et al. (20) and Xu et al. (200) developed an elegant solution to this problem, in which the low rank matrix is detered as the imizer of a convex program. In this work, we concentrate in the noisy version of robust PCA (RPCA) obtained by solving the unconstrained program L,O R m n 2 X L O 2 F + λ L + λ O, () where X R m n is the data matrix, L is a low rank matrix and O is an error matrix with a sparse number of non-zero coefficients (with arbitrarily large magnitude). L denotes the matrix nuclear norm, defined as the sum of the singular values of L. The parameters λ and λ are a positive scalar parameters controlling the rank of L and the sparsity level in the outliers, respectively, and can be set automatically as in Ramirez & Sapiro (202). This particular formulation of RPCA has attracted significant interest in the machine learning, computer vision, and signal processing communities, and was successfully used in various applications such as Wagner et al. (20); Peng et al. (200); Zhou et al. (200); Mu et al. (20). A challenge often encountered in modern settings is that the flow of new input data is permanent. Then, the

2 robust low rank model needs to be adapted constantly since the principal directions can change over time, calling for developing efficient online techniques recently proposed in Qiu & Vaswani (20); Tan et al. (20); Mateos & Giannakis (202); Balzano et al. (200). Significant amount of effort has been devoted to developing optimization algorithms for efficiently solving () and its noiseless formulation Candès et al. (20); Lin et al. (2009); Ma et al. (20); Recht & Ré (20). Despite the permanent progress reported in the literature, state-of-the-art algorithms still have prohibitive complexity and latency for real-time processing. In the sparse coding domain, a similar situation was encountered a few years ago. This motivated significant effort in the deep learning community, e.g., by Ranzato et al. (2007) and Goodfellow et al. (2009), aig at overcog this problem. Works by Jarrett et al. (2009) and Kavukcuoglu et al. (200) proposed learning non-linear regressors capable of producing good approximations of the true sparse codes in a fixed amount of time. Gregor & LeCun (200) introduced an approach in which the regressors are multilayer artificial neural networks with an architecture inspired by first order optimization algorithms for solving sparse coding problems. This work was later extended for handling more general structured sparse models by Sprechmann et al. (202a). With this motivation, in this paper we propose to extend these ideas to the RPCA context. We propose to design regressors capable of approximating online RPCA in a very fast way. To the best of our knowledge, this type of encoders have never been developed before. Our RPCA encoders are learned by imizing various objective functions that allow their use in several different contexts. In Section 2, we present our approach to RPCA and discus exact optimization algorithms to solve it. In Section 3, we introduce the new robust encoders and the new objective functions used for their training. In Section 4, we present several experimental results. Conclusions are drawn in Section 5. For further details, the reader is referred to Sprechmann et al. (202b). 2 Online RPCA via non-convex factorization Using the result from Srebro & Shraibman (2005) and Recht et al. (200), Mateos & Giannakis (202) proposed to reformulate () as U,S,O 2 X US O 2 F + λ 2 ( U 2 F + S 2 F ) + λ O, (2) with U R m q, S R q n, and O R m n. Here, q is rough upper bound on rank(l) q. This decomposition reveals a lot of structure hidden in the problem. The low rank component can now be thought of as an under-complete dictionary U, with q atoms, multiplied by a matrix S containing in its columns the corresponding coefficients for each data vector in X. This interpretation brings our problem close to that of dictionary learning in the sparse modeling domain; see Mairal et al. (2009). While this new factorized formulation drastically reduces the number of optimization variables, problem (2) is no longer convex. Fortunately, Mardani et al. (20) showed that any stationary point of (2), {U, S, O}, satisfying X US O 2 λ is an optimal solution of (2). Thus, problem (2) can be solved using an alternating imization or block coordinate scheme, in which the cost function is imized with respect to each individual optimization variable while keeping the other ones fixed, without the risk of falling into a stationary point that may not be globally optimal. This will be exploited to design our fast encoders. Robust low dimensional projections: Let us assume that we have already learned a low dimensional model, U R m q, from some data X US + O R m n. Suppose that we are given a new input vector x R m drawn from the same distribution as X. Then x can be decomposed as x Us + o, where Us represents the low dimensional component, n is a small perturbation and o is a sparse outlier vector. This can be done by solving s R q,o R m 2 x Us o λ 2 s λ o. (3) Unlike dictionary learning problems, e.g. in Mairal et al. (2009), here the columns of the dictionary U are not constrained to have unit norm. The robust low dimensional projection (3) is a convex program that can be solved using several methods. We are interested in choosing an optimization algorithm that can be further used to define the architecture of trainable encoders for simultaneously estimating s and o. With this in d, we choose to use the block-coordinate imization scheme. 2

3 The solution of (3) is given by s = (U T U λ I) U T (x t o) and o = π λ (ˆx Us), when fixing o and s respectively. Here π λ is the scalar soft-thresholding operator with parameter λ R m, which applies a soft-threshold λ i to each component of the input vector. In our case, λ = λ. Note that we can write the estimates in a recursive manner as, o k+ = π λ (b k ), b k+ = b k + W(o k+ o k ), where k is the iteration number, and W = UHU T and H = (U T U λ I). After convergence, s can be obtained as s = H(x o). Online RPCA: So far, we assumed that the entire data matrix X was available a priori. We now address the case when the data samples {x t } t N, x t R m, arrive sequentially; the index t should be understood as a temporal variable. Online RPCA aims at estimating and refining the model as the data come in. An alternating imization algorithm for solving the online counterpart of (2) goes as follows: When a new data vector x t is received, we first obtain its representation {s t, o t } as a low dimensional projection (3) given the current model estimate, U t, which can be done using strategies such as block-coordinate descent methods with warm restarts described Mairal et al. (2009), or recursive least squares Mateos & Giannakis (202). 3 Online RPCA via fast trainable encoders The main contributions of this paper is the construction of trainable regressors that can be used for approximating the solution of (3). The main idea is to build a parametric regressor z = (s, o) = h(x, Θ), with some set of parameters, collectively denoted as Θ. Thus, we need to define an architecture for h and a learning algorithm in order to detere Θ. Following the fast sparse coding methods in Gregor & LeCun (200) and Sprechmann et al. (202a), we propose to use feed-forward multi-layer architecture where each layer implements a single iteration of the block-coordinate imization scheme described in Section 2. The parameters of the network are the matrices W and H and the thresholds λ (extra flexibility is obtained by learning different thresholds λ i for each component). The encoder architecture is depicted in Sprechmann et al. (202b). Each layer essentially consists of the nonlinear thresholding operator π λ followed by a linear operation W. The network parameters are initialized to replicate a fixed number of iterations of the exact algorithm. As a learning strategy, we propose to select the set of parameters Θ that imizes the loss function, L(Θ) = n L(Θ, x j ) (4) n on a training set X = {x,..., x n }. Here, L(Θ, x j ) is a function that measures the quality of the code z j = h(x j, Θ) produced for the data point x j. The selection of the objective function L sets the type of regressor that we are going to obtain and it is application-dependent. One of the simplest choices is to use L(Θ, x j ) = z j z j, with z j = (s j, o j ) being the j-th columns of the decomposition of the data X = (x,..., x n ) into X = US + O by the exact RPCA. This essentially trains the encoder to approximate the exact solution of the RPCA problem. In many applications, the data may not completely adhere to the assumptions of the RPCA model, and the exact solution is, therefore, not necessarily the best one. This is the case, for example, in the source separation problem presented in Huang et al. (202), where RPCA gives a very good separation of the spectrally sparse singing voice (modeled as outliers) and repetitive low-rank background accompaniment, yet the obtained signals are still not equal to the true voice and background tracks. In this case, one could use a collection of clean voice and background tracks, {s j } and {o j } respectively, to supervise the training; this leads to better separation results as reported in Section 4. In a fully online setting, one does not have access either to the exact RPCA decomposition nor the true ground truth separation. In that case, we propose to directly use L(Θ, x j ) = f(x, z), where f is the objective function in the low dimensional projection (3). In this fully unsupervised case the encoders can be viewed as stand alone RPCA regressors, which brings them close in spirit to the sparse auto-encoders proposed by Goodfellow et al. (2009). Several applications of RPCA rely on the critical assumption that the given input vectors are aligned with respect to a group of geometric transformations. Translating the work by Peng Same could be done solving (3) using proximal methods instead of block-coordinate imization. j= 3

4 Figure : Robust PCA representation of several frames (top-to-bottom) from the surveillance sequence obtained using the algorithm from Lin et al. (2009) (left), and a five layer network encoder (right). Columns in each group are, leftto-right: the reconstructed frame, its lowrank approximation (background), and the sparse outlier (foreground). Method GNSDR GSNR GSAR GSIR Ideal freq. mask ADMoM RPCA Huang et al. (202) Proximal RPCA NN RPCA Untrained NN RPCA Unsupervised NN RPCA Supervised Table : Performance of audio separation methods on the MIR-K dataset. et al. (200) to our setting, we propose to use the training objective function defined in (4) with L Tr (Θ, x j, α j ) = f(t αj (x j ), h(t αj (x j ), Θ)), where T α is a parametrized operator (with a set of parameters collectively denoted as α) that applies different geometric transformation, T αj, to each training vector x j. In the online setting, when a new data vector x t arrives, we compute its robust low rank projection by imizing L Tr (Θ, x t, α t ) over a the vector α t parameterizing the geometric transformation. In the same way, as new data arrive, the transformation of all the previously seen training vectors is updated through the imization of a loss function L(Θ, α) with respect to α. The imization of a loss function L(Θ) with respect to Θ or α is performed using a stochastic gradient descent, see Gregor & LeCun (200) and Sprechmann et al. (202a) for details. 4 Experimental results Video separation: Figure shows background and foreground separation via robust PCA on the surveillance video sequence Hall of a business building taken from Li et al. (2004). The sequence consists of images of an indoor scene shot by a static camera in a mall. The scene has a nearly constant background and walking people in the foreground. We used networks with five layers and q = 5 trained to approximate the output of the exact RPCA on a subset of the frames in the sequence. The separation produced by the fast encoder is nearly identical to the output of the exact algorithm and to the output of the code from Lin et al. (2009), used as reference, while being considerably faster. Our Matlab implementation with built-in GPU acceleration executed on an NVIDIA Tesla C2070 GPU propagates a frame through a single layer of the network in merely 92µsec. This is several orders of magnitude faster than iterative solver executed on the CPU. Audio separation: We evaluate the separation performance of the proposed methods on the MIR- K dataset from Hsu & Jang (200), containing 000 clips extraced from 0 Chinese karaoke songs. The experimental settings closely followed that of Lin et al. (2009), to which the reader is referred for further details. As the evaluation criteria, we used the BSS-EVAL metrics developed in Vincent et al. (2006), which calculate the global normalized source-to-distortion ratio (GNSDR),source-to-artifacts ratio (GSAR), source-to-interference ratio (GSIR), and signal-tonoise ratio (GSNR). All networks used 20 layers with q = 25. The following training regimes were compared: untrained parameters initialized according to Section 2 (Untrained); unsupervised learning (Unsupervised); and training supervised by the clean voice and background tracks (Supervised). Table summarizes the obtained separation performance. While unsupervised training makes fast RPCA encoders on par with the exact RPCA (at a fraction of the computational complexity and latency of the latter), significant improvement is achieved by using the supervised regime. 5 Conclusion By combining ideas from structured non-convex optimization with multi-layer neural networks, we have developed a comprehensive framework for the online learning of robust low-rank representations in real time and capable of handling large scale applications. A basic implementation already achieves several order of magnitude speedups when compared to exact solvers, opening the door for practical algorithms in various applications. 4

5 References Balzano, L., Nowak, R., and Recht, B. Online identification and tracking of subspaces from highly incomplete information. In Proc. of 48th Allerton Conf., 200. Candès, E., Li, X., Ma, Y., and Wright, J. Robust principal component analysis? Journal of the ACM, 58(3), 20. Goodfellow, I., Le, Q., Saxe, A., Lee, H., and Ng, A. Y. Measuring invariances in deep networks. In In NIPS, pp Gregor, K. and LeCun, Y. Learning fast approximations of sparse coding. In ICML, pp , 200. Hsu, C.L. and Jang, J.S.R. On the improvement of singing voice separation for monaural recordings using the MIR-K dataset. IEEE Trans. on Audio, Speech, and Lang. Proc., 8(2):30 39, 200. Huang, P-S., Chen, S.D., Smaragdis, P., and Hasegawa-Johnson, M. Singing-voice separation from monaural recordings using robust principal component analysis. In ICASSP, 202. Jarrett, K., Kavukcuoglu, K., Ranzato, M.A., and LeCun, Y. What is the best multi-stage architecture for object recognition? In CVPR, pp , Kavukcuoglu, K., Ranzato, M.A., and LeCun, Y. Fast inference in sparse coding algorithms with applications to object recognition. arxiv: , 200. Li, L., Huang, W., Gu, I. Yu-Hua, and Tian, Q. Statistical modeling of complex backgrounds for foreground object detection. IEEE Trans. Image Process., 3(): , Lin, Z., Ganesh, A., Wright, J., Wu, L., Chen, M., and Ma, Y. Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix. preprint, Ma, S., Goldfarb, D., and Chen, L. Fixed point and Bregman iterative methods for matrix rank imization. Math. Program., 28(-2):32 353, 20. Mairal, J., Bach, F., Ponce, J., and Sapiro, G. Online dictionary learning for sparse coding. In ICML, pp , Mardani, M., Mateos, G., and Giannakis, G. B. Unveiling network anomalies in large-scale networks via sparsity and low rank. In Proc. of 44th Asilomar Conf. on Signals, Systems, and Computers, 20. Mateos, G. and Giannakis, G. B. Robust PCA as bilinear decomposition with outlier-sparsity regularization. IEEE Trans. on Signal Process., 60(0): , 202. Mu, Y., Dong, J., Yuan, X., and Yan, S. Accelerated low-rank visual recovery by random projection. In CVPR, pp , 20. Peng, Y., Ganesh, A., Wright, J., Xu, W., and Ma, Y. RASL: Robust alignment by sparse and low-rank decomposition for linearly correlated images. In CVPR, pp , 200. Qiu, C. and Vaswani, N. Real-time robust principal components pursuit. arxiv.org:.788, 20. Ramirez, I. and Sapiro, G. Low-rank data modeling via the imum description length principle. In ICASSP, 202. Ranzato, M., Huang, F. J., Boureau, Y-L., and LeCun, Y. Unsupervised learning of invariant feature hierarchies with applications to object recognition. In CVPR, Recht, B. and Ré, C. Parallel stochastic gradient algorithms for large-scale matrix completion. 20. Recht, B., Fazel, M., and Parrilo, P. A. Guaranteed imum-rank solutions of linear matrix equations via nuclear norm imization. SIAM Rev., 52(3):47 50, 200. Sprechmann, P., Bronstein, A. M., and Sapiro, G. Learning efficient structured sparse models. In ICML, 202a. Sprechmann, P., Bronstein, A. M., and Sapiro, G. Learning Robust Low-Rank Representations. arxiv.org: , 202b. Srebro, N. and Shraibman, A. Rank, trace-norm and max-norm. In COLT, pp , Tan, W., Cheung, G., and Ma, Y. Face recovery in conference video streag using robust principal component analysis. In ICIP, pp , 20. Vincent, E., Gribonval, R., and Févotte, C. Performance measurement in blind audio source separation. IEEE Trans. on Audio, Speech, and Lang. Proc., 4(4): , Wagner, A., Wright, J., Ganesh, A., Zhou, Z., Mobahi, H., and Ma, Y. Towards a practical face recognition system: Robust alignment and illuation via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell., 34(2): , 20. Xu, H., Caramanis, C., and Sanghavi, S. Robust PCA via outlier pursuit. In NIPS, pp Zhou, Z., Li, X., Wright, J., Candès, E. J., and Ma, Y. Stable principal component pursuit. In ISIT, pp ,

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

Robust Principal Component Analysis (RPCA)

Robust Principal Component Analysis (RPCA) Robust Principal Component Analysis (RPCA) & Matrix decomposition: into low-rank and sparse components Zhenfang Hu 2010.4.1 reference [1] Chandrasekharan, V., Sanghavi, S., Parillo, P., Wilsky, A.: Ranksparsity

More information

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary

More information

IMA Preprint Series # 2281

IMA Preprint Series # 2281 DICTIONARY LEARNING AND SPARSE CODING FOR UNSUPERVISED CLUSTERING By Pablo Sprechmann and Guillermo Sapiro IMA Preprint Series # 2281 ( September 2009 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS UNIVERSITY

More information

ONLINE ROBUST PRINCIPAL COMPONENT ANALYSIS FOR BACKGROUND SUBTRACTION: A SYSTEM EVALUATION ON TOYOTA CAR DATA XINGQIAN XU THESIS

ONLINE ROBUST PRINCIPAL COMPONENT ANALYSIS FOR BACKGROUND SUBTRACTION: A SYSTEM EVALUATION ON TOYOTA CAR DATA XINGQIAN XU THESIS ONLINE ROBUST PRINCIPAL COMPONENT ANALYSIS FOR BACKGROUND SUBTRACTION: A SYSTEM EVALUATION ON TOYOTA CAR DATA BY XINGQIAN XU THESIS Submitted in partial fulfillment of the requirements for the degree of

More information

Limitations of Matrix Completion via Trace Norm Minimization

Limitations of Matrix Completion via Trace Norm Minimization Limitations of Matrix Completion via Trace Norm Minimization ABSTRACT Xiaoxiao Shi Computer Science Department University of Illinois at Chicago xiaoxiao@cs.uic.edu In recent years, compressive sensing

More information

FAST PRINCIPAL COMPONENT PURSUIT VIA ALTERNATING MINIMIZATION

FAST PRINCIPAL COMPONENT PURSUIT VIA ALTERNATING MINIMIZATION FAST PRICIPAL COMPOET PURSUIT VIA ALTERATIG MIIMIZATIO Paul Rodríguez Department of Electrical Engineering Pontificia Universidad Católica del Perú Lima, Peru Brendt Wohlberg T-5 Applied Mathematics and

More information

Unsupervised learning in Vision

Unsupervised learning in Vision Chapter 7 Unsupervised learning in Vision The fields of Computer Vision and Machine Learning complement each other in a very natural way: the aim of the former is to extract useful information from visual

More information

COMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS. Jingyu Yang, Xuemeng Yang, Xinchen Ye

COMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS. Jingyu Yang, Xuemeng Yang, Xinchen Ye COMPLETION OF STRUCTURALLY-INCOMPLETE MATRICES WITH REWEIGHTED LOW-RANK AND SPARSITY PRIORS Jingyu Yang, Xuemeng Yang, Xinchen Ye School of Electronic Information Engineering, Tianjin University Building

More information

Online PLCA for Real-time Semi-supervised Source Separation

Online PLCA for Real-time Semi-supervised Source Separation Online PLCA for Real-time Semi-supervised Source Separation Zhiyao Duan 1, Gautham J. Mysore 2 and Paris Smaragdis 2,3 1 EECS Department, Northwestern University, 2 Advanced Technology Labs, Adobe Systems

More information

Low Rank Representation Theories, Algorithms, Applications. 林宙辰 北京大学 May 3, 2012

Low Rank Representation Theories, Algorithms, Applications. 林宙辰 北京大学 May 3, 2012 Low Rank Representation Theories, Algorithms, Applications 林宙辰 北京大学 May 3, 2012 Outline Low Rank Representation Some Theoretical Analysis Solution by LADM Applications Generalizations Conclusions Sparse

More information

Learning based face hallucination techniques: A survey

Learning based face hallucination techniques: A survey Vol. 3 (2014-15) pp. 37-45. : A survey Premitha Premnath K Department of Computer Science & Engineering Vidya Academy of Science & Technology Thrissur - 680501, Kerala, India (email: premithakpnath@gmail.com)

More information

Real-time Background Subtraction via L1 Norm Tensor Decomposition

Real-time Background Subtraction via L1 Norm Tensor Decomposition Real-time Background Subtraction via L1 Norm Tensor Decomposition Taehyeon Kim and Yoonsik Choe Yonsei University, Seoul, Korea E-mail: pyomu@yonsei.ac.kr Tel/Fax: +82-10-2702-7671 Yonsei University, Seoul,

More information

IN some applications, we need to decompose a given matrix

IN some applications, we need to decompose a given matrix 1 Recovery of Sparse and Low Rank Components of Matrices Using Iterative Method with Adaptive Thresholding Nematollah Zarmehi, Student Member, IEEE and Farokh Marvasti, Senior Member, IEEE arxiv:1703.03722v2

More information

arxiv: v1 [cs.cv] 9 Sep 2013

arxiv: v1 [cs.cv] 9 Sep 2013 Learning Transformations for Clustering and Classification arxiv:139.274v1 [cs.cv] 9 Sep 213 Qiang Qiu Department of Electrical and Computer Engineering Duke University Durham, NC 2778, USA Guillermo Sapiro

More information

ELEG Compressive Sensing and Sparse Signal Representations

ELEG Compressive Sensing and Sparse Signal Representations ELEG 867 - Compressive Sensing and Sparse Signal Representations Gonzalo R. Arce Depart. of Electrical and Computer Engineering University of Delaware Fall 211 Compressive Sensing G. Arce Fall, 211 1 /

More information

Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection

Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection Direct Matrix Factorization and Alignment Refinement: Application to Defect Detection Zhen Qin (University of California, Riverside) Peter van Beek & Xu Chen (SHARP Labs of America, Camas, WA) 2015/8/30

More information

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1

Section 5 Convex Optimisation 1. W. Dai (IC) EE4.66 Data Proc. Convex Optimisation page 5-1 Section 5 Convex Optimisation 1 W. Dai (IC) EE4.66 Data Proc. Convex Optimisation 1 2018 page 5-1 Convex Combination Denition 5.1 A convex combination is a linear combination of points where all coecients

More information

Blockwise Matrix Completion for Image Colorization

Blockwise Matrix Completion for Image Colorization Blockwise Matrix Completion for Image Colorization Paper #731 Abstract Image colorization is a process of recovering the whole color from a monochrome image given that a portion of labeled color pixels.

More information

Real-Time Principal Component Pursuit

Real-Time Principal Component Pursuit Real-Time Principal Component Pursuit Graeme Pope, Manuel Baumann, Christoph Studer, and Giuseppe Durisi Dept. of Information Technology and Electrical Engineering, ETH Zurich, Zurich, Switzerland e-mails:

More information

Novel Lossy Compression Algorithms with Stacked Autoencoders

Novel Lossy Compression Algorithms with Stacked Autoencoders Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is

More information

Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification?

Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification? Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification? Roberto Rigamonti, Matthew A. Brown, Vincent Lepetit CVLab, EPFL Lausanne, Switzerland firstname.lastname@epfl.ch

More information

Outlier Pursuit: Robust PCA and Collaborative Filtering

Outlier Pursuit: Robust PCA and Collaborative Filtering Outlier Pursuit: Robust PCA and Collaborative Filtering Huan Xu Dept. of Mechanical Engineering & Dept. of Mathematics National University of Singapore Joint w/ Constantine Caramanis, Yudong Chen, Sujay

More information

An Approach for Reduction of Rain Streaks from a Single Image

An Approach for Reduction of Rain Streaks from a Single Image An Approach for Reduction of Rain Streaks from a Single Image Vijayakumar Majjagi 1, Netravati U M 2 1 4 th Semester, M. Tech, Digital Electronics, Department of Electronics and Communication G M Institute

More information

Robust Transfer Principal Component Analysis with Rank Constraints

Robust Transfer Principal Component Analysis with Rank Constraints Robust Transfer Principal Component Analysis with Rank Constraints Yuhong Guo Department of Computer and Information Sciences Temple University, Philadelphia, PA 19122, USA yuhong@temple.edu Abstract Principal

More information

arxiv: v1 [cs.cv] 1 Aug 2013

arxiv: v1 [cs.cv] 1 Aug 2013 Domain-invariant Face Recognition using Learned Low-rank Transformation arxiv:1308.0275v1 [cs.cv] 1 Aug 2013 Abstract Qiang Qiu Duke University Durham, NC, 27708 qiang.qiu@duke.edu Guillermo Sapiro Duke

More information

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro Learning Low-rank Transformations: Algorithms and Applications Qiang Qiu Guillermo Sapiro Motivation Outline Low-rank transform - algorithms and theories Applications Subspace clustering Classification

More information

Online Subspace Estimation and Tracking from Missing or Corrupted Data

Online Subspace Estimation and Tracking from Missing or Corrupted Data Online Subspace Estimation and Tracking from Missing or Corrupted Data Laura Balzano www.ece.wisc.edu/~sunbeam Work with Benjamin Recht and Robert Nowak Subspace Representations Capture Dependencies Subspace

More information

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN

Face Recognition Using Vector Quantization Histogram and Support Vector Machine Classifier Rong-sheng LI, Fei-fei LEE *, Yan YAN and Qiu CHEN 2016 International Conference on Artificial Intelligence: Techniques and Applications (AITA 2016) ISBN: 978-1-60595-389-2 Face Recognition Using Vector Quantization Histogram and Support Vector Machine

More information

ELEG Compressive Sensing and Sparse Signal Representations

ELEG Compressive Sensing and Sparse Signal Representations ELEG 867 - Compressive Sensing and Sparse Signal Representations Introduction to Matrix Completion and Robust PCA Gonzalo Garateguy Depart. of Electrical and Computer Engineering University of Delaware

More information

Photometric Stereo with Auto-Radiometric Calibration

Photometric Stereo with Auto-Radiometric Calibration Photometric Stereo with Auto-Radiometric Calibration Wiennat Mongkulmann Takahiro Okabe Yoichi Sato Institute of Industrial Science, The University of Tokyo {wiennat,takahiro,ysato} @iis.u-tokyo.ac.jp

More information

Sparse Coding for Learning Interpretable Spatio-Temporal Primitives

Sparse Coding for Learning Interpretable Spatio-Temporal Primitives Sparse Coding for Learning Interpretable Spatio-Temporal Primitives Taehwan Kim TTI Chicago taehwan@ttic.edu Gregory Shakhnarovich TTI Chicago gregory@ttic.edu Raquel Urtasun TTI Chicago rurtasun@ttic.edu

More information

Sparse Coding and Dictionary Learning for Image Analysis

Sparse Coding and Dictionary Learning for Image Analysis Sparse Coding and Dictionary Learning for Image Analysis Part IV: Recent Advances in Computer Vision and New Models Francis Bach, Julien Mairal, Jean Ponce and Guillermo Sapiro CVPR 10 tutorial, San Francisco,

More information

LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA

LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES Qiang Qiu and Guillermo Sapiro Duke University, Durham, NC 2778, USA ABSTRACT Learning a transformation-based dimension reduction, thereby compressive,

More information

arxiv: v1 [eess.sp] 1 Nov 2017

arxiv: v1 [eess.sp] 1 Nov 2017 LEARNED CONVOLUTIONAL SPARSE CODING Hillel Sreter Raja Giryes School of Electrical Engineering, Tel Aviv University, Tel Aviv, Israel, {hillelsr@mail., raja@}tau.ac.il arxiv:7.00328v [eess.sp] Nov 207

More information

Sparse Shape Registration for Occluded Facial Feature Localization

Sparse Shape Registration for Occluded Facial Feature Localization Shape Registration for Occluded Facial Feature Localization Fei Yang, Junzhou Huang and Dimitris Metaxas Abstract This paper proposes a sparsity driven shape registration method for occluded facial feature

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques

Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques Patch-Based Color Image Denoising using efficient Pixel-Wise Weighting Techniques Syed Gilani Pasha Assistant Professor, Dept. of ECE, School of Engineering, Central University of Karnataka, Gulbarga,

More information

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING Jianzhou Feng Li Song Xiaog Huo Xiaokang Yang Wenjun Zhang Shanghai Digital Media Processing Transmission Key Lab, Shanghai Jiaotong University

More information

MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo

MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo MULTI-POSE FACE HALLUCINATION VIA NEIGHBOR EMBEDDING FOR FACIAL COMPONENTS Yanghao Li, Jiaying Liu, Wenhan Yang, Zongg Guo Institute of Computer Science and Technology, Peking University, Beijing, P.R.China,

More information

Fault detection with principal component pursuit method

Fault detection with principal component pursuit method Journal of Physics: Conference Series PAPER OPEN ACCESS Fault detection with principal component pursuit method Recent citations - Robust Principal Component Pursuit for Fault Detection in a Blast Furnace

More information

Effectiveness of Sparse Features: An Application of Sparse PCA

Effectiveness of Sparse Features: An Application of Sparse PCA 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Convex Optimization / Homework 2, due Oct 3

Convex Optimization / Homework 2, due Oct 3 Convex Optimization 0-725/36-725 Homework 2, due Oct 3 Instructions: You must complete Problems 3 and either Problem 4 or Problem 5 (your choice between the two) When you submit the homework, upload a

More information

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING

ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING ADAPTIVE LOW RANK AND SPARSE DECOMPOSITION OF VIDEO USING COMPRESSIVE SENSING Fei Yang 1 Hong Jiang 2 Zuowei Shen 3 Wei Deng 4 Dimitris Metaxas 1 1 Rutgers University 2 Bell Labs 3 National University

More information

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all

More information

Weighted-CS for reconstruction of highly under-sampled dynamic MRI sequences

Weighted-CS for reconstruction of highly under-sampled dynamic MRI sequences Weighted- for reconstruction of highly under-sampled dynamic MRI sequences Dornoosh Zonoobi and Ashraf A. Kassim Dept. Electrical and Computer Engineering National University of Singapore, Singapore E-mail:

More information

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016 CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:

More information

A Geometric Analysis of Subspace Clustering with Outliers

A Geometric Analysis of Subspace Clustering with Outliers A Geometric Analysis of Subspace Clustering with Outliers Mahdi Soltanolkotabi and Emmanuel Candés Stanford University Fundamental Tool in Data Mining : PCA Fundamental Tool in Data Mining : PCA Subspace

More information

Sparsity and image processing

Sparsity and image processing Sparsity and image processing Aurélie Boisbunon INRIA-SAM, AYIN March 6, Why sparsity? Main advantages Dimensionality reduction Fast computation Better interpretability Image processing pattern recognition

More information

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari Laboratory for Advanced Brain Signal Processing Laboratory for Mathematical

More information

NTHU Rain Removal Project

NTHU Rain Removal Project People NTHU Rain Removal Project Networked Video Lab, National Tsing Hua University, Hsinchu, Taiwan Li-Wei Kang, Institute of Information Science, Academia Sinica, Taipei, Taiwan Chia-Wen Lin *, Department

More information

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N.

ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N. ADVANCED IMAGE PROCESSING METHODS FOR ULTRASONIC NDE RESEARCH C. H. Chen, University of Massachusetts Dartmouth, N. Dartmouth, MA USA Abstract: The significant progress in ultrasonic NDE systems has now

More information

Modeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations

Modeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations Modeling Visual Cortex V4 in Naturalistic Conditions with Invariant and Sparse Image Representations Bin Yu Departments of Statistics and EECS University of California at Berkeley Rutgers University, May

More information

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction Ham Rara, Shireen Elhabian, Asem Ali University of Louisville Louisville, KY {hmrara01,syelha01,amali003}@louisville.edu Mike Miller,

More information

Image Restoration and Background Separation Using Sparse Representation Framework

Image Restoration and Background Separation Using Sparse Representation Framework Image Restoration and Background Separation Using Sparse Representation Framework Liu, Shikun Abstract In this paper, we introduce patch-based PCA denoising and k-svd dictionary learning method for the

More information

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong

Proceedings of the International MultiConference of Engineers and Computer Scientists 2018 Vol I IMECS 2018, March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong , March 14-16, 2018, Hong Kong TABLE I CLASSIFICATION ACCURACY OF DIFFERENT PRE-TRAINED MODELS ON THE TEST DATA

More information

Supervised Translation-Invariant Sparse Coding

Supervised Translation-Invariant Sparse Coding Supervised Translation-Invariant Sparse Coding Jianchao Yang,KaiYu, Thomas Huang Beckman Institute, University of Illinois at Urbana-Champaign NEC Laboratories America, Inc., Cupertino, California {jyang29,

More information

Consistent dictionary learning for signal declipping

Consistent dictionary learning for signal declipping Consistent dictionary learning for signal declipping Lucas Rencker 1 ( ), Francis Bach, Wenwu Wang 1, and Mark D. Plumbley 1 1 Centre for Vision, Speech and Signal Processing, University of Surrey, Guildford,

More information

OR-PCA with MRF for Robust Foreground Detection in Highly Dynamic Backgrounds

OR-PCA with MRF for Robust Foreground Detection in Highly Dynamic Backgrounds OR-PCA with MRF for Robust Foreground Detection in Highly Dynamic Backgrounds Sajid Javed 1, Seon Ho Oh 1, Andrews Sobral 2, Thierry Bouwmans 2 and Soon Ki Jung 1 1 School of Computer Science and Engineering,

More information

Extended Dictionary Learning : Convolutional and Multiple Feature Spaces

Extended Dictionary Learning : Convolutional and Multiple Feature Spaces Extended Dictionary Learning : Convolutional and Multiple Feature Spaces Konstantina Fotiadou, Greg Tsagkatakis & Panagiotis Tsakalides kfot@ics.forth.gr, greg@ics.forth.gr, tsakalid@ics.forth.gr ICS-

More information

Stacked Denoising Autoencoders for Face Pose Normalization

Stacked Denoising Autoencoders for Face Pose Normalization Stacked Denoising Autoencoders for Face Pose Normalization Yoonseop Kang 1, Kang-Tae Lee 2,JihyunEun 2, Sung Eun Park 2 and Seungjin Choi 1 1 Department of Computer Science and Engineering Pohang University

More information

Robust and Secure Iris Recognition

Robust and Secure Iris Recognition Robust and Secure Iris Recognition Vishal M. Patel University of Maryland, UMIACS pvishalm@umiacs.umd.edu IJCB 2011 Tutorial Sparse Representation and Low-Rank Representation for Biometrics Outline Iris

More information

Change-Point Estimation of High-Dimensional Streaming Data via Sketching

Change-Point Estimation of High-Dimensional Streaming Data via Sketching Change-Point Estimation of High-Dimensional Streaming Data via Sketching Yuejie Chi and Yihong Wu Electrical and Computer Engineering, The Ohio State University, Columbus, OH Electrical and Computer Engineering,

More information

CLEANING UP TOXIC WASTE: REMOVING NEFARIOUS CONTRIBUTIONS TO RECOMMENDATION SYSTEMS

CLEANING UP TOXIC WASTE: REMOVING NEFARIOUS CONTRIBUTIONS TO RECOMMENDATION SYSTEMS CLEANING UP TOXIC WASTE: REMOVING NEFARIOUS CONTRIBUTIONS TO RECOMMENDATION SYSTEMS Adam Charles, Ali Ahmed, Aditya Joshi, Stephen Conover, Christopher Turnes, Mark Davenport Georgia Institute of Technology

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

Convex optimization algorithms for sparse and low-rank representations

Convex optimization algorithms for sparse and low-rank representations Convex optimization algorithms for sparse and low-rank representations Lieven Vandenberghe, Hsiao-Han Chao (UCLA) ECC 2013 Tutorial Session Sparse and low-rank representation methods in control, estimation,

More information

Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks

Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Show, Discriminate, and Tell: A Discriminatory Image Captioning Model with Deep Neural Networks Zelun Luo Department of Computer Science Stanford University zelunluo@stanford.edu Te-Lin Wu Department of

More information

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude

Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude Advanced phase retrieval: maximum likelihood technique with sparse regularization of phase and amplitude A. Migukin *, V. atkovnik and J. Astola Department of Signal Processing, Tampere University of Technology,

More information

Discriminative sparse model and dictionary learning for object category recognition

Discriminative sparse model and dictionary learning for object category recognition Discriative sparse model and dictionary learning for object category recognition Xiao Deng and Donghui Wang Institute of Artificial Intelligence, Zhejiang University Hangzhou, China, 31007 {yellowxiao,dhwang}@zju.edu.cn

More information

Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision 2013 IEEE International Conference on Computer Vision Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Tae-Hyun Oh* KAIST Hyeongwoo Kim KAIST Yu-Wing Tai KAIST Jean-Charles Bazin

More information

Seismic data reconstruction with Generative Adversarial Networks

Seismic data reconstruction with Generative Adversarial Networks Seismic data reconstruction with Generative Adversarial Networks Ali Siahkoohi 1, Rajiv Kumar 1,2 and Felix J. Herrmann 2 1 Seismic Laboratory for Imaging and Modeling (SLIM), The University of British

More information

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING Pan Ji, Yiran Zhong, Hongdong Li, Mathieu Salzmann, Australian National University, Canberra NICTA, Canberra {pan.ji,hongdong.li}@anu.edu.au,mathieu.salzmann@nicta.com.au

More information

The Benefit of Tree Sparsity in Accelerated MRI

The Benefit of Tree Sparsity in Accelerated MRI The Benefit of Tree Sparsity in Accelerated MRI Chen Chen and Junzhou Huang Department of Computer Science and Engineering, The University of Texas at Arlington, TX, USA 76019 Abstract. The wavelet coefficients

More information

Generalized Lasso based Approximation of Sparse Coding for Visual Recognition

Generalized Lasso based Approximation of Sparse Coding for Visual Recognition Generalized Lasso based Approximation of Sparse Coding for Visual Recognition Nobuyuki Morioka The University of New South Wales & NICTA Sydney, Australia nmorioka@cse.unsw.edu.au Shin ichi Satoh National

More information

Sparse Models in Image Understanding And Computer Vision

Sparse Models in Image Understanding And Computer Vision Sparse Models in Image Understanding And Computer Vision Jayaraman J. Thiagarajan Arizona State University Collaborators Prof. Andreas Spanias Karthikeyan Natesan Ramamurthy Sparsity Sparsity of a vector

More information

Visual object classification by sparse convolutional neural networks

Visual object classification by sparse convolutional neural networks Visual object classification by sparse convolutional neural networks Alexander Gepperth 1 1- Ruhr-Universität Bochum - Institute for Neural Dynamics Universitätsstraße 150, 44801 Bochum - Germany Abstract.

More information

Slides adapted from Marshall Tappen and Bryan Russell. Algorithms in Nature. Non-negative matrix factorization

Slides adapted from Marshall Tappen and Bryan Russell. Algorithms in Nature. Non-negative matrix factorization Slides adapted from Marshall Tappen and Bryan Russell Algorithms in Nature Non-negative matrix factorization Dimensionality Reduction The curse of dimensionality: Too many features makes it difficult to

More information

Sparsity Based Regularization

Sparsity Based Regularization 9.520: Statistical Learning Theory and Applications March 8th, 200 Sparsity Based Regularization Lecturer: Lorenzo Rosasco Scribe: Ioannis Gkioulekas Introduction In previous lectures, we saw how regularization

More information

An efficient algorithm for sparse PCA

An efficient algorithm for sparse PCA An efficient algorithm for sparse PCA Yunlong He Georgia Institute of Technology School of Mathematics heyunlong@gatech.edu Renato D.C. Monteiro Georgia Institute of Technology School of Industrial & System

More information

arxiv: v1 [cs.lg] 20 Dec 2013

arxiv: v1 [cs.lg] 20 Dec 2013 Unsupervised Feature Learning by Deep Sparse Coding Yunlong He Koray Kavukcuoglu Yun Wang Arthur Szlam Yanjun Qi arxiv:1312.5783v1 [cs.lg] 20 Dec 2013 Abstract In this paper, we propose a new unsupervised

More information

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction

Akarsh Pokkunuru EECS Department Contractive Auto-Encoders: Explicit Invariance During Feature Extraction Akarsh Pokkunuru EECS Department 03-16-2017 Contractive Auto-Encoders: Explicit Invariance During Feature Extraction 1 AGENDA Introduction to Auto-encoders Types of Auto-encoders Analysis of different

More information

arxiv: v1 [cs.cv] 28 Sep 2018

arxiv: v1 [cs.cv] 28 Sep 2018 Camera Pose Estimation from Sequence of Calibrated Images arxiv:1809.11066v1 [cs.cv] 28 Sep 2018 Jacek Komorowski 1 and Przemyslaw Rokita 2 1 Maria Curie-Sklodowska University, Institute of Computer Science,

More information

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Sakrapee Paisitkriangkrai, Chunhua Shen, Anton van den Hengel The University of Adelaide,

More information

Uniqueness in Bilinear Inverse Problems with Applications to Subspace and Sparsity Models

Uniqueness in Bilinear Inverse Problems with Applications to Subspace and Sparsity Models Uniqueness in Bilinear Inverse Problems with Applications to Subspace and Sparsity Models Yanjun Li Joint work with Kiryung Lee and Yoram Bresler Coordinated Science Laboratory Department of Electrical

More information

Combining Mask Estimates for Single Channel Audio Source Separation using Deep Neural Networks

Combining Mask Estimates for Single Channel Audio Source Separation using Deep Neural Networks Combining Mask Estimates for Single Channel Audio Source Separation using Deep Neural Networks Emad M. Grais, Gerard Roma, Andrew J.R. Simpson, Mark D. Plumbley Centre for Vision, Speech and Signal Processing,

More information

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies

P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies P257 Transform-domain Sparsity Regularization in Reconstruction of Channelized Facies. azemi* (University of Alberta) & H.R. Siahkoohi (University of Tehran) SUMMARY Petrophysical reservoir properties,

More information

INTERACTIVE REFINEMENT OF SUPERVISED AND SEMI-SUPERVISED SOUND SOURCE SEPARATION ESTIMATES

INTERACTIVE REFINEMENT OF SUPERVISED AND SEMI-SUPERVISED SOUND SOURCE SEPARATION ESTIMATES INTERACTIVE REFINEMENT OF SUPERVISED AND SEMI-SUPERVISED SOUND SOURCE SEPARATION ESTIMATES Nicholas J. Bryan Gautham J. Mysore Center for Computer Research in Music and Acoustics, Stanford University Adobe

More information

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model

Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model Illumination-Robust Face Recognition based on Gabor Feature Face Intrinsic Identity PCA Model TAE IN SEOL*, SUN-TAE CHUNG*, SUNHO KI**, SEONGWON CHO**, YUN-KWANG HONG*** *School of Electronic Engineering

More information

Semi-supervised Data Representation via Affinity Graph Learning

Semi-supervised Data Representation via Affinity Graph Learning 1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073

More information

Exploring Curve Fitting for Fingers in Egocentric Images

Exploring Curve Fitting for Fingers in Egocentric Images Exploring Curve Fitting for Fingers in Egocentric Images Akanksha Saran Robotics Institute, Carnegie Mellon University 16-811: Math Fundamentals for Robotics Final Project Report Email: asaran@andrew.cmu.edu

More information

Blind Identification of Graph Filters with Multiple Sparse Inputs

Blind Identification of Graph Filters with Multiple Sparse Inputs Blind Identification of Graph Filters with Multiple Sparse Inputs Santiago Segarra, Antonio G. Marques, Gonzalo Mateos & Alejandro Ribeiro Dept. of Electrical and Systems Engineering University of Pennsylvania

More information

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation

Module 1 Lecture Notes 2. Optimization Problem and Model Formulation Optimization Methods: Introduction and Basic concepts 1 Module 1 Lecture Notes 2 Optimization Problem and Model Formulation Introduction In the previous lecture we studied the evolution of optimization

More information

EE795: Computer Vision and Intelligent Systems

EE795: Computer Vision and Intelligent Systems EE795: Computer Vision and Intelligent Systems Spring 2012 TTh 17:30-18:45 FDH 204 Lecture 11 140311 http://www.ee.unlv.edu/~b1morris/ecg795/ 2 Outline Motion Analysis Motivation Differential Motion Optical

More information

Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal

Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal Sparse Component Analysis (SCA) in Random-valued and Salt and Pepper Noise Removal Hadi. Zayyani, Seyyedmajid. Valliollahzadeh Sharif University of Technology zayyani000@yahoo.com, valliollahzadeh@yahoo.com

More information

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2 Vol.3, Issue 3, 2015, Page.1115-1021 Effect of Anti-Forensics and Dic.TV Method for Reducing Artifact in JPEG Decompression 1 Deepthy Mohan, 2 Sreejith.H 1 PG Scholar, 2 Assistant Professor Department

More information

Conditional gradient algorithms for machine learning

Conditional gradient algorithms for machine learning 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

A low rank based seismic data interpolation via frequencypatches transform and low rank space projection

A low rank based seismic data interpolation via frequencypatches transform and low rank space projection A low rank based seismic data interpolation via frequencypatches transform and low rank space projection Zhengsheng Yao, Mike Galbraith and Randy Kolesar Schlumberger Summary We propose a new algorithm

More information

BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo

BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo 2014 IEEE International Conference on Acoustic, Speech and Signal Processing (ICASSP) BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS Yongqin Zhang, Jiaying Liu, Mading Li, Zongming

More information

Deep Generative Models Variational Autoencoders

Deep Generative Models Variational Autoencoders Deep Generative Models Variational Autoencoders Sudeshna Sarkar 5 April 2017 Generative Nets Generative models that represent probability distributions over multiple variables in some way. Directed Generative

More information

Optimization. Industrial AI Lab.

Optimization. Industrial AI Lab. Optimization Industrial AI Lab. Optimization An important tool in 1) Engineering problem solving and 2) Decision science People optimize Nature optimizes 2 Optimization People optimize (source: http://nautil.us/blog/to-save-drowning-people-ask-yourself-what-would-light-do)

More information