Discriminative sparse model and dictionary learning for object category recognition

Similar documents
Sparse Coding and Dictionary Learning for Image Analysis

String distance for automatic image classification

IMA Preprint Series # 2281

Image Restoration and Background Separation Using Sparse Representation Framework

Sparsity and image processing

INCOHERENT DICTIONARY LEARNING FOR SPARSE REPRESENTATION BASED IMAGE DENOISING

Department of Electronics and Communication KMP College of Engineering, Perumbavoor, Kerala, India 1 2

Greedy algorithms for Sparse Dictionary Learning

Generalized Lasso based Approximation of Sparse Coding for Visual Recognition

Greedy vs. L1 Convex Optimization in Sparse Coding Ren, Huamin; Pan, Hong; Olsen, Søren Ingvor; Moeslund, Thomas B.

Efficient Implementation of the K-SVD Algorithm and the Batch-OMP Method

Effectiveness of Sparse Features: An Application of Sparse PCA

SHIP WAKE DETECTION FOR SAR IMAGES WITH COMPLEX BACKGROUNDS BASED ON MORPHOLOGICAL DICTIONARY LEARNING

Multi-polarimetric SAR image compression based on sparse representation

SPARSE COMPONENT ANALYSIS FOR BLIND SOURCE SEPARATION WITH LESS SENSORS THAN SOURCES. Yuanqing Li, Andrzej Cichocki and Shun-ichi Amari

Efficient Implementation of the K-SVD Algorithm using Batch Orthogonal Matching Pursuit

Discriminative Representative Selection via Structure Sparsity

Aggregating Descriptors with Local Gaussian Metrics

QUATERNION-BASED SPARSE REPRESENTATION OF COLOR IMAGE. Licheng Yu, Yi Xu, Hongteng Xu, Hao Zhang

BSIK-SVD: A DICTIONARY-LEARNING ALGORITHM FOR BLOCK-SPARSE REPRESENTATIONS. Yongqin Zhang, Jiaying Liu, Mading Li, Zongming Guo

Learning-based Methods in Vision

Improving Recognition through Object Sub-categorization

An Approach for Reduction of Rain Streaks from a Single Image

Comparison of Optimization Methods for L1-regularized Logistic Regression

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Cervigram Image Segmentation Based On Reconstructive Sparse Representations

IMAGE RESTORATION VIA EFFICIENT GAUSSIAN MIXTURE MODEL LEARNING

Modified Iterative Method for Recovery of Sparse Multiple Measurement Problems

A Learned Dictionary Model for Texture Classification

DALM-SVD: ACCELERATED SPARSE CODING THROUGH SINGULAR VALUE DECOMPOSITION OF THE DICTIONARY

Hierarchical Matching Pursuit for Image Classification: Architecture and Fast Algorithms

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

arxiv: v1 [cs.cv] 2 Sep 2013

Multi-Class Image Classification: Sparsity Does It Better

REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY. Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin

IMAGE SUPER-RESOLUTION BASED ON DICTIONARY LEARNING AND ANCHORED NEIGHBORHOOD REGRESSION WITH MUTUAL INCOHERENCE

Action Recognition By Learnt Class-Specific Overcomplete Dictionaries

Sparse Models in Image Understanding And Computer Vision

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization

Image Denoising via Group Sparse Eigenvectors of Graph Laplacian

BLOCKWISE COORDINATE DESCENT SCHEMES FOR SPARSE REPRESENTATION

Hyperspectral Data Classification via Sparse Representation in Homotopy

Bilevel Sparse Coding

Efficient Kernels for Identifying Unbounded-Order Spatial Features

Combining Selective Search Segmentation and Random Forest for Image Classification

Patch-based Image Denoising: Probability Distribution Estimation vs. Sparsity Prior

Image Denoising Using Sparse Representations

Beyond bags of features: Adding spatial information. Many slides adapted from Fei-Fei Li, Rob Fergus, and Antonio Torralba

NTHU Rain Removal Project

Research Article A Modified Sparse Representation Method for Facial Expression Recognition

Sparsity Based Regularization

PATCH-DISAGREEMENT AS A WAY TO IMPROVE K-SVD DENOISING

Dictionary of gray-level 3D patches for action recognition

Robust Face Recognition via Sparse Representation

Previously. Part-based and local feature models for generic object recognition. Bag-of-words model 4/20/2011

Multiresponse Sparse Regression with Application to Multidimensional Scaling

Ecole normale supérieure, Paris Guillermo Sapiro and Andrew Zisserman

Modeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations

Virtual Training Samples and CRC based Test Sample Reconstruction and Face Recognition Experiments Wei HUANG and Li-ming MIAO

Restoring STM images via Sparse Coding: noise and artifact removal

Improving Image Segmentation Quality Via Graph Theory

ANALYSIS SPARSE CODING MODELS FOR IMAGE-BASED CLASSIFICATION. Sumit Shekhar, Vishal M. Patel and Rama Chellappa

Pattern recognition (3)

Extended Dictionary Learning : Convolutional and Multiple Feature Spaces

IMAGE RETRIEVAL USING VLAD WITH MULTIPLE FEATURES

LETTER Learning Co-occurrence of Local Spatial Strokes for Robust Character Recognition

Visual Tracking by Dictionary Learning and Motion Estimation

arxiv: v1 [cs.lg] 20 Dec 2013

Bag of Words Models. CS4670 / 5670: Computer Vision Noah Snavely. Bag-of-words models 11/26/2013

Sparse representation of terrains for procedural modeling

Sparse Coding and Mid-Level Superpixel-Feature for l0-graph Based Unsupervised Image Segmentation

Learning Convolutional Feature Hierarchies for Visual Recognition

Sparse learned representations for image restoration

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

Frequent Inner-Class Approach: A Semi-supervised Learning Technique for One-shot Learning

Sparse Coding based Frequency Adaptive Loop Filtering for Video Coding

Multipath Sparse Coding Using Hierarchical Matching Pursuit

Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling

An efficient algorithm for sparse PCA

Supervised Translation-Invariant Sparse Coding

INPAINTING COLOR IMAGES IN LEARNED DICTIONARY. Bijenička cesta 54, 10002, Zagreb, Croatia

arxiv: v3 [cs.cv] 3 Oct 2012

Multipath Sparse Coding Using Hierarchical Matching Pursuit

Discovering Visual Hierarchy through Unsupervised Learning Haider Razvi

CS6670: Computer Vision

An R Package flare for High Dimensional Linear Regression and Precision Matrix Estimation

Ranking Error-Correcting Output Codes for Class Retrieval

Multipath Sparse Coding Using Hierarchical Matching Pursuit

Sparse Coding for Learning Interpretable Spatio-Temporal Primitives

Robust Multimodal Dictionary Learning

Extracting Spatio-temporal Local Features Considering Consecutiveness of Motions

Face recognition based on improved BP neural network

Beyond Bags of Features

Non-local Similarity Regularized Sparsity. Model for Hyperspectral Target Detection

Spatial Latent Dirichlet Allocation

A GRAPH-CUT APPROACH TO IMAGE SEGMENTATION USING AN AFFINITY GRAPH BASED ON l 0 SPARSE REPRESENTATION OF FEATURES

Graph Regularized Sparse Coding by Modified Online Dictionary Learning

Dictionary Learning in Texture Classification

EFFICIENT DICTIONARY LEARNING IMPLEMENTATION ON THE GPU USING OPENCL

SHAPE-INCLUDED LABEL-CONSISTENT DISCRIMINATIVE DICTIONARY LEARNING: AN APPROACH TO DETECT AND SEGMENT MULTI-CLASS OBJECTS IN IMAGES

Transcription:

Discriative sparse model and dictionary learning for object category recognition Xiao Deng and Donghui Wang Institute of Artificial Intelligence, Zhejiang University Hangzhou, China, 31007 {yellowxiao,dhwang}@zju.edu.cn Abstract Recent researches have well established that sparse signal models have led to outstanding performances in signal, image and video processing tasks. This success is mainly due to the fact that natural signals such as images admit sparse representations of some redundant basis, also called dictionary. This paper focuses on learning discriative dictionaries instead of reconstructive ones. It has been shown that discriative dictionaries, which are composed of sparse reconstruction and class discriation terms, outperform reconstructive ones for image classification tasks. Experimental results in image classification tasks using examples from the Caltech 101 Object Categories show that the proposed method is efficient and can achieve a higher recognition rate than reconstructive methods. Keywords Sparse representation, dictionary learning, object category recognition, classification I. INTRODUCTION Category-level object recognition deals with one kind of visual tasks where all instances of a category are to be recognized. The key challenge in this problem consists in how to learn representative models of visual object categories for recognition. The learned object models are required not only to capture common semantic information of certain category, but also to generate exactly discriative capability. In this paper, we focus on this challenging problem and adapt a discriative sparse model and dictionary learning method for object category recognition. An important representative model for category-level object recognition is the bag of words (BoW) approach which has been investigated by several studies[1], [], [3]. The key idea behind this model is to learn a visual codebook (SIFT descriptors) for each category by using an unsupervised clustering algorithm such as k-means. One major problem with this approach is that the construction of visual codebook and classifier training are two disconnected stage. That says the visual codebook of a category is learned separately without using the information of other categories. This reduces the discriative ability of learned codebooks. Another competitive model for category-level object recognition is sparse signal representation[4]. In this model, the sparsity is confirmed as a powerful prior for visual inference and the semantic information of the image is naturally encoded as sparse representation with respect to predefined bases or learned dictionaries. The key problem in this model is how to correctly choose the basis or dictionary for representing Correspondence author. image information. Using predefined bases such as curvelets, wedgelets, wavelets and its variants[1], isn t suitable for specific vision tasks because these bases don t use the semantic prior of given sample images. Learning a task-specific dictionaries instead can lead to more significant improvements than using predefined ones[4]. In current sparse representation and dictionary learning methods, the main goal of learning a dictionary for sparse representation is to reconstruct image information instead of category recognition. To recognize an object category needs to combine sparse reconstruction and class discriation. This motivates us to learn a discriative sparse model for representing common semantic information of specific category as well as for discriating different object categories. To achieve this purpose, we follow the strategy proposed in[5] for local image analysis. N dictionaries for N categories are learned simultaneously under the constraint of reconstructive and discriative loss function. The reconstruction errors of learned dictionaries on image patches are used to derive an object category recognition. The paper is organized as follows. Section presents background of sparse representation and dictionary learning. A discriative sparse model and dictionary learning method for object category recognition is introduced in Section 3. Section 4 presents experimental results on object categories recognition. Concluding remarks are given in Section 5. II. BACKGROUND AND RELATED WORK In classical sparse coding problem, one considers a signal y R n and a dictionary D = [d 1,..., d k ] R n k. Under the assumption that a signal can be approximately represented as a sparse linear combination of the atoms from the dictionary, y can be approximately represented as Dα for some sparse coefficient vector α R k. Thus finding the sparse representation of y leads to the following optimization problem: α p, s.t. y Dα α R k ε (1) where the most frequently-used values for p are 0 and 1. However, when p is 0, the solution of this l 0 norm optimization problem has been proven to be an NP-hard problem. Thus, approximate solutions are considered, and several algorithms such as Matching pursuit (MP)[7] and Orthogonal Matching pursuit (OMP)[6] have been proposed in the past decades. When p is 1, this l 1 norm optimization problem is the

well-known Basic Pursuit[9] or Lasso[8] problem. Several algorithms such as LARS[10] have been proposed to solve this problem efficiently. Recent study has demonstrated that using learned dictionaries rather than predefined ones can lead to state-of-the-art performances[4]. In practice, overlapping patches instead of the whole images are sparsely decomposed because the natural images are usually very large. For an image I, suppose there are M overlapping patches {y i } M i=1 Rn of size n from the image I. Then the dictionary D R n k is learning via solving the following optimization problem: D,A M α i p, s.t. y i Dα i ε () i=1 where A = [α 1,..., α M ] is a matrix in R k M, y i is the i-th patch of the image I written as a column vector, α i is the corresponding sparse code, Dα i is the estimate of y i and the dictionary D is normalized with each column having an unit l norm. Several algorithms have been proposed for addressing the above optimization problem (). Among these algorithms, the K-SVD algorithm has been widely used for reconstruction tasks in recent years[11]. III. DISCRIMINATIVE SPARSE MODEL In our task, some sample images are provided as training set. They are classified in advance by object categories. We need to learn the representative models of object categories as prior knowledge for recognizing the category of test image. As introduced above, the current sparse model provides an effective framework to extract common and compact representation of one category by dictionary learning algorithm. Such dictionaries are learned separately for each category and obtain optimal decomposition basis for image reconstruction. For object category recognition, the reconstruction residual of test image on learned dictionary of each category is used as a criteria for discriation. The flow of this method is shown in Figure 1. It s worth notice that dictionary learning of one category don t use any information of other categories which may be an important cue for discriation. Thus, we present a discriative sparse model and dictionary learning algorithm to seek optimal representative basis for object category recognition. Consider C sets of signals {{y j i }n j i=1 }C j=1 Rn N, where N denotes the total number of signals. Our goal is to learn C discriative dictionaries {D j } C j=1 Rn kc, making dictionary D j corresponding to the j-th class good at reconstructing this class while bad for the other classes simultaneously. Following the strategy in [5], we make some improvements. In our framework, the residual y Dα is used as a criteria for discriation. To achieve the goal of discriation, we introduce a discriative function, which is the log function of the reciprocal of the softmax function f j (x) = C j (x 1,..., x C ) = log( e (x l x j ) ) (3) l=1 Thus, finding the class label of testing sets is equivalent to imize (3). Then, adding this discriative term into the original dictionary updating formulation, we have {D j} C j=1 i=1 C i ({ y l D j α lj }C j=1 )+γ y l D i α li ) ( n i l=1 (4) where γ is the parameter governing the trade-off between reconstruction and discriation, and α lj denotes the corresponding sparse coefficients of signal y l on the dictionary D j which have been computed during the sparse coding step. Solving this optimization problem is hard because it is neither convex nor differentiable. Thus, truncated Newton method is considered to give an approximate formulation easier to solve. Then, updating the p-th dictionary is equivalent to solve the following formulation n i D R n k i=1 l=1 w l y l Dα lp, where w l = C i x p ({ y l D j α lj }C j=1 ) + γ1 p(i) x p denotes y l D p α lp, 1 p(i) is 1 if i = p and 0 otherwise. Inspired by the idea of the K-SVD algorithm, (5) can be rewritten as d j,α j w l i=1 l {1,...,n i} ω l (y l d p α p ) d j α j where ω l = {i 1 i N, s.t. α l (i) 0} denotes indices of signals that use the atom d j, d p and α p denote all the d j and α j whose indices are not p. The matrix form of the above formulation is (Y 1 DA)W (Ej = d j α j )W 1 d j,α j F d j,α j F (7) where Y = [y 1,..., y N ] R n N is the signal matrix, A = [α 1,..., α N ] R k N is the coefficient matrix, and W = diag(w 1,..., w N ) is the diagonal matrix of the weights w l. Using the definition of condensed SVD decomposition, matrix E j W 1 can be written as UΛV T. Then, the solution for atom d j is u 1, which is the eigenvector associated to the largest eigenvalue of E j W 1, and the solution for the non-zero coefficient α j is ((W 1 ) ( 1) λ 1 v 1 ) T. The details are shown in Algorithm 1. Suppose we have obtained C dictionaries D i, one for each class, via the above discriative K-SVD algorithm. The strategy for classification in our paper is based on the reconstructive errors, which are generated by approximating each patch using the C dictionaries D i. Thus, finding the right class label of patch y is equivalent to solve (5) (6) i y D i α i (8) IV. EXPERIMENTAL RESULTS In this section, we study whether our proposed discriative framework can enhance the discriative power of dictionaries and is more efficient when applied to object category recognition tasks. Meanwhile, we also study the roles

(a) Dictionary learning stage. (b) Object category recognition stage. Fig. 1. The flow of our method for object category recognition. TABLE I E RROR Class bike none Fig.. The left two images are bike and ketch from the Caltech 101 Object categories dataset. The right two images are the corresponding patches extracted from the natural images of bike and ketch. The upper one has 300 random patches of size 0 0. The below one has 100 random patches of size 3 3. of the three parameters in our proposed framework size of patches, size of dictionaries, and γ via our experiments. We have selected 50, 000 patches which are centered and have normalized ℓ norm, respectively from images of different classes from the Caltech 101 Object Categories dataset. There are two different sizes of patches which are 0 0 and 3 3 in our experiments. Figure shows the partial original patches of these two different sizes. We then divide patches of each category into a training set and a testing set, respectively of 15, 000 and 35, 000 patches. A. Choices of parameters Size of patches. In this experiment, we study the role of size of patches in our proposed framework. We use two groups of patches of size 0 0 and 3 3 from natural images of bike and background. The size of dictionaries k is 144, the sparsity factor L is 4, and γ is 3.0. Then 15 iterations of our framework are processed. Figure 3 shows the dictionaries learned from the two groups of patches of the two sizes. The RATES OF DICTIONARIES OF DIFFERENT PARAMETERS. p IS THE SIZE OF PATCHES, AND k IS THE SIZE OF DICTIONARIES. p = 400,k = 144 p = 104,k = 144 1% p = 104, k = 64 error rates of these two groups of dictionaries are showed in Table I. Comparing the dictionaries of bike with those of background, we can see that there are more arc-shaped patches in the dictionaries of bike while in the dictionaries of background there are more bar-shaped patches. Those arc-shaped patches correspond to the wheels of bikes while in the images of background there are rarely things of round or other shapes. Meanwhile, comparing the dictionaries of patches of two sizes, we observe that the dictionary of bike learned from patches of larger size has more arc-shaped patches than that learned from patches of smaller size and dictionaries of larger patches are more distinguishable than those of smaller patches. Moreover, from the error rates showed in Table I we observe that dictionaries of larger patches are more discriative than those of smaller patches. Patches of small size contain more reconstructive information, however, lose important information that distinguishes one category from others. Thus, when learning dictionaries applied for discriation tasks, patches of large size are selected. Size of dictionaries. This experiment studies the role of size of dictionaries in our framework. We use patches of size 3 3 from natural images of bike and background, but this time we change the size of dictionaries k into 64 and γ remains unchanged. Figure 3 shows the dictionaries learned in this experiment and the error rates of this experiment are

Algorithm 1 The Algorithm of discriative sparse model Input: N signals y l R n which belong to C different classes. Output: C dictionaries D i R n k. Initialization: Initialize the C dictionaries D i with l normalized columns. Loop: Repeat until convergence (or stopping criterion): (1) Sparse coding: For C dictionaries D i and N vectors y l finding the solution of the following optimization problem: y α l D i α li, s.t. α li p L li () Dictionary updating: For i = 1,,..., C, for j = 1,,..., k, update d j, the j-th atom of D i : Select the set of signals using atom d j : ω j = {l 1,,..., N α li [j] 0} If there are no indices in set ω j For l = 1,,..., N, find the least represented signal y via solving the following maximization problem: y = arg max y l y l D i α li Update the j-th atom d j using the normalized signal y, and the corresponding non-zero coefficients using zero vector. Else Compute the representation error matrix E: E = Y p j d p α p Restrict E to the columns included in the set ω j, and obtain E j. Compute the weights w l in ω j. Update a new d and the associated nonzero coefficients α li via solving the following optimization problem: (E j d j α j )W 1 d,α li F The new d is the eigenvector associated to the largest eigenvalue of E j W 1, W = diag(w1,..., w N ) showed in Table I. Comparing dictionaries learned in this experiment with those of size 144 which are learned under the same conditions showed in Figure 3, we observe that dictionaries learned in the last group of experiments are more distinguishable from each other than those learned this time. Meanwhile, from the results showed in Table I we can see that the discriation power of dictionaries of smaller size is weaker than that of dictionaries of larger size. If the size of dictionaries is too small, information in learned dictionaries is not sufficient for discriation. Thus, the size of dictionaries should be large enough for helping dictionaries more discriative when applied them for object category recognition tasks. Parameters γ. In this group of experiments, we study the role of parameters γ in our framewrok, meanwhile explore the role of λ introduced in Mairal s paper[5]. We use patches Fig. 3. The upper and lower images are dictionaries of bike and background respectively. The left two are dictionaries of size 144 learned from patches of size 0 0. The middle two are dictionaries of size 144 learned from patches of size 3 3. The right two are dictionaries of size 64 learned from patches of size 3 3. of larger size in the first group of experiments, and the other parameters remain the same as our first group of experiments except the values of parameters γ and λ. In these experiments, we use different values of parameters γ = {1, 1.5, 3, 5, 7.5, 10, 1, 15, 0, 50}, and λ = {0.1, 0., 0.5, 0.75, 1, 1.5,, 3, 8, 10}. Figure 4 shows the residuals of the training patches on learned dictionaries of bike and background. The percentage error rates when γ = 1 for any λ of bike and background are {6.5, 5, 57.5, 77.5, 50, 55, 7.5, 30, 60, 47.5} and {35, 57.5, 60, 57.5, 67.5, 7.5, 67.5, 5.5, 5.5, 67.5} and the error rates of other γ and λ are close and low. From Figure 4, we observe that when the values of parameter γ are larger than some value, the residuals remain roughly the same no mater what the values of γ and λ are. Thus, we omit parameter λ in our framework. The reconstructive parameter γ controls the trade-off between reconstruction and discriation. (4) has solution if and only if the diagonal matrix W of the weights w l is positive definite, and from (5) we can easily see that γ plays an important role computing the weights w l. Thus, the value of γ should be large enough to ensure the existence of the solution of our proposed discriative formulation. Meanwhile, if γ is too large, the discriative power will reduce according to (4). Thus, the value of γ should small enough to ensure the discriative ability. B. Object category recognition This subsection studies whether our proposed framework is more efficient than other methods of object categories recognition and also efficient for multiclass recognition tasks. In our first group of experiments, we have mainly compared our method with the original reconstructive dictionary learning and the BoW methods. We use the patches of size 0 0, and the size of dictionaries is 64. We want to study whether our method has an advantage over the reconstructive one when the parameters are good for reconstruction while bad for discriation. 15 iterations are performed in our framework and the classical dictionary learning framework. Figure 5 shows the dictionaries learned via these two methods. The discriation results are showed in Table II. From results

Fig. 7. left: dictionary of butterfly. Middle: dictionary of ketch. Right: dictionary of face. TABLE III E RROR Image class error rates Fig. 4. The left two images show the residuals of background and bike projected on the background dictionary. The right two are images of residuals of them projected on the bike dictionary. Fig. 5. The left two images are dictionaries of bike and background learned via our framework. The right two show dictionaries learned by the classical dictionary learning framework. showed in Table II, we can easily see that the discriation power of our method is more efficient than that of the classical dictionary learning and BoW methods. This group of experiments shows that our proposed framework is also efficient when applied to multicalss recognition tasks. We use patches of size 3 3 from natural images of butterfly, ketch, and face. The size of dictionaries k is 144, the sparsity factor L is 4, and γ is 3. 0 iterations of our framework are performed. Figure 6 shows the partial natural images used in this experiment. Figure 7 shows the dictionaries learned in this experiment and Table III shows the error rates of these three classes. TABLE II E RROR RATES OF THREE METHODS. Image class bike background Fig. 6. discriation 1% reconstruction 1 Partial natural images used in this experiment. BoW RATES OF THREE CLASSES. butterfly 10% ketch face V. C ONCLUSION This paper proposed a discriative sparse models and dictionary learning method for object category recognition. Experiments with the Caltech 101 Object Categories shows that our framework is efficient when applied for object category recognition tasks. Meanwhile, we also find out the roles of the four parameters size of patches, size of dictionaries, and γ via comparing these experiments. ACKNOWLEDGMENTS This work is supported by 973 Program (Project No. 010CB37905) and Natural Science Foundations of China (No.6107118). R EFERENCES [1] Gabriella Csurka and Christopher R. Dance and Lixin Fan and Jutta Willamowski and Cdric Bray, Visual categorization with bags of keypoints, In Workshop on Statistical Learning in Computer Vision, ECCV, 1-, 004. [] Li Fei-fei and P. Perona, A bayesian hierarchical model for learning natural scene categories, CVPR, 54-531, 005. [3] J. Winn and A. Criisi and T. Minka, Object categorization by learned universal visual dictionary, ICCV, 1800-1807, 005. [4] Wright, J. and Ma, Y. and Mairal, J. and Sapiro, G. and Huang, T.S. and Yan, S., Sparse Representation for Computer Vision and Pattern Recognition, PIEEE, 1031-1044, 010. [5] Mairal, Julien and Bach, Francis and Ponce, Jean and Sapiro, Guillermo and Zisserman, Andrew, Discriative learned dictionaries for local image analysis, Computer Vision and Pattern Recognition, CVPR, 1-8, 008. [6] Y. C. Pati and R. Rezaiifar and Y. C. Pati R. Rezaiifar and P. S. Krishnaprasad, Orthogonal Matching Pursuit: Recursive Function Approximation with Applications to Wavelet Decomposition, Proceedings of the 7 th Annual Asilomar Conference on Signals, Systems, and Computers, 40-44, 1993. [7] Stphane Mallat and Zhifeng Zhang, Matching Pursuit With TimeFrequency Dictionaries, IEEE Transactions on Signal Processing, 33973415, 1993. [8] Robert Tibshirani, Regression Shrinkage and Selection Via the Lasso, Journal of the Royal Statistical Society, Series B, 67-88, 1994. [9] Chen, Scott Shaobing and Donoho, David L. and Saunders, Michael A., Atomic Decomposition by Basis Pursuit, SIAM J. Sci. Comput., 33-61, 1998. [10] Efron, Bradley and Hastie, Trevor and Johnstone, Iain and Tibshirani, Robert, Least angle regression, Ann. Statist., 407-499, 004. [11] Aharon, M. and Elad, M. and Bruckstein, A., K-SVD: An Algorithm for Designing Overcomplete Dictionaries for Sparse Representation, Signal Processing, IEEE Transactions on [see also Acoustics, Speech, and Signal Processing, IEEE Transactions on], 4311 43, 006. [1] Mallat, Stephane, A Wavelet Tour of Signal Processing, 3rd ed., Third Edition: The Sparse Way, Academic Press, 3 ed, 008.