Projective dictionary pair learning for pattern classification

Size: px
Start display at page:

Download "Projective dictionary pair learning for pattern classification"

Transcription

1 Projective dictionary pair learning for pattern classification Shuhang Gu 1, Lei Zhang 1, Wangmeng Zuo 2, Xiangchu Feng 3 1 Dept. of Computing, The Hong Kong Polytechnic University, Hong Kong, China 2 School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China 3 Dept. of Applied Mathematics, Xidian University, Xi an, China {cssgu, cslzhang}@comp.polyu.edu.h cswmzuo@gmail.com, xcfeng@mail.xidian.edu.cn Abstract Discriminative dictionary learning (DL) has been widely studied in various pattern classification problems. Most of the existing DL methods aim to learn a synthesis dictionary to represent the input signal while enforcing the representation coefficients and/or representation residual to be discriminative. However, the l 0 or l 1 -norm sparsity constraint on the representation coefficients adopted in most DL methods maes the training and testing phases time consuming. We propose a new discriminative DL framewor, namely projective dictionary pair learning (DPL), which learns a synthesis dictionary and an analysis dictionary jointly to achieve the goal of signal representation and discrimination. Compared with conventional DL methods, the proposed DPL method can not only greatly reduce the time complexity in the training and testing phases, but also lead to very competitive accuracies in a variety of visual classification tass. 1 Introduction Sparse representation represents a signal as the linear combination of a small number of atoms chosen out of a dictionary, and it has achieved a big success in various image processing and computer vision applications [1, 2]. The dictionary plays an important role in the signal representation process [3]. By using a predefined analytical dictionary (e.g., wavelet dictionary, Gabor dictionary) to represent a signal, the representation coefficients can be produced by simple inner product operations. Such a fast and explicit coding maes analytical dictionary very attractive in image representation; however, it is less effective to model the complex local structures of natural images. Sparse representation with a synthesis dictionary has been widely studied in recent years [2, 4, 5]. With synthesis dictionary, the representation coefficients of a signal are usually obtained via an l p -norm (p 1) sparse coding process, which is computationally more expensive than analytical dictionary based representation. However, synthesis based sparse representation can better model the complex image local structures and it has led to many state-of-the-art results in image restoration [6]. Another important advantage lies in that the synthesis based sparse representation model allows us to easily learn a desired dictionary from the training data. The seminal wor of KSVD [1] tells us that an over-complete dictionary can be learned from example natural images, and it can lead to much better image reconstruction results than the analytically designed off-the-shelf dictionaries. Inspired by KSVD, many dictionary learning (DL) methods have been proposed and achieved stateof-the-art performance in image restoration tass. The success of DL in image restoration problems triggers its applications in image classification tass. Different from image restoration, assigning the correct class label to the test sample is the goal of classification problems; therefore, the discrimination capability of the learned dictionary is 1

2 of the major concern. To this end, supervised dictionary learning methods have been proposed to promote the discriminative power of the learned dictionary [4, 5, 7, 8, 9]. By encoding the query sample over the learned dictionary, both the coding coefficients and the coding residual can be used for classification, depending on the employed DL model. Discriminative DL has led to many stateof-the-art results in pattern recognition problems. One popular strategy of discriminative DL is to learn a shared dictionary for all classes while enforcing the coding coefficients to be discriminative [4, 5, 7]. A classifier on the coding coefficients can be trained simultaneously to perform classification. Mairal et al. [7] proposed to learn a dictionary and a corresponding linear classifier in the coding vector space. In the label consistent KSVD (LC-KSVD) method, Jiang et al. [5] introduced a binary class label sparse code matrix to encourage samples from the same class to have similar sparse codes. In [4], Mairal et al. proposed a tas driven dictionary learning (TDDL) framewor, which minimizes different ris functions of the coding coefficients for different tass. Another popular line of research in DL attempts to learn a structured dictionary to promote discrimination between classes [2, 8, 9, 10]. The atoms in the structured dictionary have class labels, and the class-specific representation residual can be computed for classification. Ramirez et al. [8] introduced an incoherence promotion term to encourage the sub-dictionaries of different classes to be independent. Yang et al. [9] proposed a Fisher discrimination dictionary learning (FDDL) method which applies the Fisher criterion to both representation residual and representation coefficient. Wang et al. [10] proposed a max-margin dictionary learning (MMDL) algorithm from the large margin perspective. In most of the existing DL methods, l 0 -norm or l 1 -norm is used to regularize the representation coefficients since sparser coefficients are more liely to produce better classification results. Hence a sparse coding step is generally involved in the iterative DL process. Although numerous algorithms have been proposed to improve the efficiency of sparse coding [11, 12], the use of l 0 -norm or l 1 - norm sparsity regularization is still a big computation burden and maes the training and testing inefficient. It is interesting to investigate whether we can learn discriminative dictionaries but without the costly l 0 -norm or l 1 -norm sparsity regularization. In particular, it would be very attractive if the representation coefficients can be obtained by linear projection instead of nonlinear sparse coding. To this end, in this paper we propose a projective dictionary pair learning (DPL) framewor to learn a synthesis dictionary and an analysis dictionary jointly for pattern classification. The analysis dictionary is trained to generate discriminative codes by efficient linear projection, while the synthesis dictionary is trained to achieve class-specific discriminative reconstruction. The idea of using functions to predict the representation coefficients is not new, and fast approximate sparse coding methods have been proposed to train nonlinear functions to generate sparse codes [13, 14]. However, there are clear difference between our DPL model and these methods. First, in DPL the synthesis dictionary and analysis dictionary are trained jointly, which ensures that the representation coefficients can be approximated by a simple linear projection function. Second, DPL utilizes class label information and promotes discriminative power of the representation codes. One related wor to this paper is the analysis-based sparse representation prior learning [15, 16], which represents a signal from a dual viewpoint of the commonly used synthesis model. Analysis prior learning tries to learn a group of analysis operators which have sparse responses to the latent clean signal. Sprechmann et al. [17] proposed to train a group of analysis operators for classification; however, in the testing phase a costly sparsity-constrained optimization problem is still required. Feng et al. [18] jointly trained a dimensionality reduction transform and a dictionary for face recognition. The discriminative dictionary is trained in the transformed space, and sparse coding is needed in both the training and testing phases. The contribution of our wor is two-fold. First, we introduce a new DL framewor, which extends the conventional discriminative synthesis dictionary learning to discriminative synthesis and analysis dictionary pair learning (DPL). Second, the DPL utilizes an analytical coding mechanism and it largely improves the efficiency in both the training and testing phases. Our experiments in various visual classification datasets show that DPL achieves very competitive accuracy with state-of-the-art DL algorithms, while it is significantly faster in both training and testing. 2

3 2 Projective Dictionary Pair Learning 2.1 Discriminative dictionary learning Denote by X = [X 1,..., X,..., X K ] a set of p-dimensional training samples from K classes, where X R p n is the training sample set of class, and n is the number of samples of each class. Discriminative DL methods aim to learn an effective data representation model from X for classification tass by exploiting the class label information of training data. Most of the state-of-the-art discriminative DL methods [5, 7, 9] can be formulated under the following framewor: min D,A X DA 2 F +λ A p +Ψ(D, A, Y), (1) where λ 0 is a scalar constant, Y represents the class label matrix of samples in X, D is the synthesis dictionary to be learned, and A is the coding coefficient matrix of X over D. In the training model (1), the data fidelity term X DA 2 F ensures the representation ability of D; A p is the l p -norm regularizer on A; and Ψ(D, A, Y) stands for some discrimination promotion function, which ensures the discrimination power of D and A. As we introduced in Section 1, some DL methods [4, 5, 7] learn a shared dictionary for all classes and a classifier on the coding coefficients simultaneously, while some DL methods [8, 9, 10] learn a structured dictionary to promote discrimination between classes. However, they all employ l 0 or l 1 -norm sparsity regularizer on the coding coefficients, maing the training stage and the consequent testing stage inefficient. In this wor, we extend the conventional DL model in (1), which learns a discriminative synthesis dictionary, to a novel DPL model, which learns a pair of synthesis and analysis dictionaries. No costly l 0 or l 1 -norm sparsity regularizer is required in the proposed DPL model, and the coding coefficients can be explicitly obtained by linear projection. Fortunately, DPL does not sacrifice the classification accuracy while achieving significant improvement in the efficiency, as demonstrated by our extensive experiments in Section The dictionary pair learning model The conventional discriminative DL model in (1) aims to learn a synthesis dictionary D to sparsely represent the signal X, and a costly l 1 -norm sparse coding process is needed to resolve the code A. Suppose that if we can find an analysis dictionary, denoted by P R mk p, such that the code A can be analytically obtained as A = PX, then the representation of X would become very efficient. Based on this idea, we propose to learn such an analysis dictionary P together with the synthesis dictionary D, leading to the following DPL model: {P,D }=arg min P,D X DPX 2 F +Ψ(D, P, X, Y), (2) where Ψ(D, P, X, Y) is some discrimination function. D and P form a dictionary pair: the analysis dictionary P is used to analytically code X, and the synthesis dictionary D is used to reconstruct X. The discrimination power of the DPL model depends on the suitable design of Ψ(D, P, X, Y). We propose to learn a structured synthesis dictionary D = [D 1,..., D,..., D K ] and a structured analysis dictionary P = [P 1 ;... ; P ;... ; P K ], where {D R p m, P R m p } forms a subdictionary pair corresponding to class. Recent studies on sparse subspace clustering [19] have proved that a sample can be represented by its corresponding dictionary if the signals satisfy certain incoherence condition. With the structured analysis dictionary P, we want that the sub-dictionary P can project the samples from class i, i, to a nearly null space, i.e., P X i 0, i. (3) Clearly, with (3) the coefficient matrix PX will be nearly bloc diagonal. On the other hand, with the structured synthesis dictionary D, we want that the sub-dictionary D can well reconstruct the data matrix X from its projective code matrix P X ; that is, the dictionary pair should minimize the reconstruction error: min X D P X 2 F. (4) P,D =1 Based on the above analysis, we can readily have the following DPL model: {P, D } = arg min P,D =1 X D P X 2 F +λ P X 2 F, s.t. d i (5) 3

4 Algorithm 1 Discriminative synthesis&analysis dictionary pair learning (DPL) Input: Training samples for K classes X = [X 1, X 2,..., X K], parameter λ, τ, m; 1: Initialize D (0) and P (0) as random matrixes with unit Frobenious norm, t = 0; 2: while not converge do 3: t t + 1; 4: for i=1:k do 5: Update A (t) by (8); 6: Update P (t) by (10); 7: Update D (t) by (12); 8: end for 9: end while Output: Analysis dictionary P, synthesis dictionary D. where X denotes the complementary data matrix of X in the whole training set X, λ > 0 is a scalar constant, and d i denotes the ith atom of synthesis dictionary D. We constrain the energy of each atom d i in order to avoid the trivial solution of P = 0 and mae the DPL more stable. The DPL model in (5) is not a sparse representation model, while it enforces group sparsity on the code matrix PX (i.e., PX is nearly bloc diagonal). Actually, the role of sparse coding in classification is still an open problem, and some researchers argued that sparse coding may not be crucial to classification tass [20, 21]. Our findings in this wor are supportive to this argument. The D- PL model leads to very competitive classification performance with those sparse coding based DL models, but it is much faster. 2.3 Optimization The objective function in (5) is generally non-convex. We introduce a variable matrix A and relax (5) to the following problem: K {P, A, D }=arg min X i D A 2 F +τ P X A 2 F +λ P X 2 F, s.t. d i (6) P,A,D =1 where τ is a scalar constant. All terms in the above objective function are characterized by Frobenius norm, and (6) can be easily solved. We initialize the analysis dictionary P and synthesis dictionary D as random matrices with unit Frobenius norm, and then alternatively update A and {D, P}. The minimization can be alternated between the following two steps. (1) Fix D and P, update A A = arg min A =1 X D A 2 F +τ P X A 2 F. (7) This is a standard least squares problem and we have the closed-form solution: A = (D T D + τi) 1 (τp X + D T X ). (8) (2) Fix A, update D and P: { P =arg min P =1 τ P X A 2 F +λ P X 2 F ; D =arg min D =1 X D A 2 F, s.t. d i The closed-form solutions of P can be obtained as: P = τa X T (τx X T + λ X XT + γi) 1, (10) where γ = 10e 4 is a small number. The D problem can be optimized by introducing a variable S: min D,S =1 X D A 2 F, s.t. D = S, s i (11) The optimal solution of (11) can be obtained by the ADMM algorithm: D (r+1) =arg min D =1 X D A 2 F +ρ D S (r) + T (r) 2 F, S (r+1) =arg min S =1 ρ D(r+1) S + T (r) 2 F, s.t. s i 2 2 1, T (r+1) =T (r) + D (r+1) S (r+1), update ρ if appropriate. (9) (12) 4

5 (a) (a) P Py y Py 2 2 (b) (b) (b) y yd y P DDy P Py y * * 2 2 * * 2 * * 2 Figure 1: (a) The representation codes and (b) reconstruction error on the Extended YaleB dataset. In each step of optimization, we have closed form solutions for variables A and P, and the ADMM based optimization of D converges rapidly. The training of the proposed DPL model is much faster than most of previous discriminative DL methods. The proposed DPL algorithm is summarized in Algorithm 1. When the difference between the energy in two adjacent iterations is less than 0.01, the iteration stops. The analysis dictionary P and the synthesis dictionary D are then output for classification. One can see that the first sub-objective function in (9) is a discriminative analysis dictionary learner, focusing on promoting the discriminative power of P; the second sub-objective function in (9) is a representative synthesis dictionary learner, aiming to minimize the reconstruction error of the input signal with the coding coefficients generated by the analysis dictionary P. When the minimization process converges, a balance between the discrimination and representation power of the model can be achieved. 2.4 Classification scheme In the DPL model, the analysis sub-dictionary P is trained to produce small coefficients for samples from classes other than, and it can only generate significant coding coefficients for samples from class. Meanwhile, the synthesis sub-dictionary D is trained to reconstruct the samples of class from their projective coefficients P X ; that is, the residual X D P X 2 F will be small. On the other hand, since P X i, i, will be small and D is not trained to reconstruct X i, the residual X i D P X i 2 F will be much larger than X D P X 2 F. In the testing phase, if the query sample y is from class, its projective coding vector by P (i.e., P y ) will be more liely to be significant, while its projective coding vectors by P i, i, tend to be small. Consequently, the reconstruction residual y D P y 2 2 tends to be much smaller than the residuals y D i P i y 2 2, i. Let us use the Extended YaleB face dataset [22] to illustrate this. (The detailed experimental setting can be found in Section 3.) Fig. 1(a) shows the l 2 -norm of the coefficients P y, where the horizontal axis refers to the index of y and the vertical axis refers to the index of P. One can clearly see that P y 2 2 has a nearly bloc diagonal structure, and the diagonal blocs are produced by the query samples which have the same class labels as P. Fig. 1(b) shows the reconstruction residual y D P y 2 2. One can see that y D P y 2 2 also has a bloc diagonal structure, and only the diagonal blocs have small residuals. Clearly, the classspecific reconstruction residual can be used to identify the class label of y, and we can naturally have the following classifier associated with the DPL model: 2.5 Complexity and Convergence identity(y) = arg min i y D i P i y 2. (13) Complexity In the training phase of DPL, A, P and D are updated alternatively. In each iteration, the time complexities of updating A, P and D are O(mpn + m 3 + m 2 n), O(mnp + p 3 + mp 2 ) and O(W(pmn + m 3 + m 2 p + p 2 m)), respectively, where W is the iteration number in ADMM algorithm for updating D. We experimentally found that in most cases W is less than 20. In many applications, the number of training samples and the number of dictionary atoms for each class are much smaller than the dimension p. Thus the major computational burden in the training phase of DPL is on updating P, which involves an inverse of a p p matrix {τx X T + λ X X T + γi}. Fortunately, this 2 2 5

6 Energy Iteration Number Figure 2: The convergence curve of DPL on the AR database. matrix will not change in the iteration, and thus the inverse of it can be pre-computed. This greatly accelerates the training process. In the testing phase, our classification scheme is very efficient. The computation of class-specific reconstruction error y D P y 2 only has a complexity of O(mp). Thus, the total complexity of our model to classify a query sample is O(Kmp). Convergence The objective function in (6) is a bi-convex problem for {(D, P), (A)}, e.g., by fixing A the function is convex for D and P, and by fixing D and P the function is convex for A. The convergence of such a problem has already been intensively studied [23], and the proposed optimization algorithm is actually an alternate convex search (ACS) algorithm. Since we have the optimal solutions of updating A, P and D, and our objective function has a general lower bound 0, our algorithm is guaranteed to converge to a stationary point. A detailed convergence analysis can be found in our supplementary file. It is empirically found that the proposed DPL algorithm converges rapidly. Fig. 2 shows the convergence curve of our algorithm on the AR face dataset [24]. One can see that the energy drops quicly and becomes very small after 10 iterations. In most of our experiments, our algorithm will converge in less than 20 iterations. 3 Experimental Results We evaluate the proposed DPL method on various visual classification datasets, including two face databases (Extended YaleB [22] and AR [24]), one object categorization database (Caltech101) [25], and one action recognition database (UCF 50 action [26]). These datasets are widely used in previous wors [5, 9] to evaluate the DL algorithms. Besides the classification accuracy, we also report the training and testing time of competing algorithms in the experiments. All the competing algorithms are implemented in Matlab except for SVM which is implemented in C. All experiments are run on a destop PC with 3.5GHz Intel CPU and 8 GB memory. The testing time is calculated in terms of the average processing time to classify a single query sample. 3.1 Parameter setting There are three parameters, m, λ and τ in the proposed DPL model. To achieve the best performance, in face recognition and object recognition experiments, we set the number of dictionary atoms as its maximum (i.e., the number of training samples) for all competing DL algorithms, including the proposed DPL. In the action recognition experiment, since the samples per class is relatively big, we set the number of dictionary atoms of each class as 50 for all the DL algorithms. Parameter τ is an algorithm parameter, and the regularization parameter λ is to control the discriminative property of P. In all the experiments, we choose λ and τ by 10-fold cross validation on each dataset. For all the competing methods, we tune their parameters for the best performance. 3.2 Competing methods We compare the proposed DPL method with the following methods: the base-line nearest subspace classifier (NSC) and linear support vector machine (SVM), sparse representation based classification (SRC) [2] and collaborative representation based classification (CRC) [21], and the state-of-the-art DL algorithms DLSI [8], FDDL [9] and LC-KSVD [5]. The original DLSI represents the test sample by each class-specific sub-dictionary. The results in [9] have shown that by coding the test sample collaboratively over the whole dictionary, the classification performance can be greatly improved. 6

7 (a) Figure 3: Sample images in the (a) Extended YaleB and (b) AR databases. Therefore, we follow the use of DLDI in [9] and denote this method as DLSI(C). For the two variants of LC-KSVD proposed in [5], we adopt the LC-KSVD2 since it can always produce better classification accuracy. 3.3 Face recognition We first evaluate our algorithm on two widely used face datasets: Extended YaleB [22] and AR [24]. The Extended YaleB database has large variations in illumination and expressions, as illustrated in Fig. 3(a). The AR database (a) involves many variations such as illumination, expressions and sunglass and scarf occlusion, as illustrated in Fig. 3(b). We follow the experimental settings in [5] for fair comparison with state-of-the-arts. A set of 2,414 face images of 38 persons are extracted from the Extended YaleB database. We randomly select half of the images per subject for training and the other half for testing. For the AR database, a set of 2,600 images of 50 female and 50 male subjects are extracted. 20 images of each subject are used for training and the remain 6 images are used for testing. We use the features provided by Jiang et al. [5] to represent the face image. The feature dimension is 504 for Extended YaleB and 540 for AR. The parameter τ is set to 0.05 on both the datasets and λ is set to 3e-3 and 5e-3 on the Extended YaleB and AR datasets, respectively. In these two experiments, we also compare with the max-margin dictionary learning (MMDL) [10] algorithm, whose recognition accuracy is cropped from the original paper but the training/testing time is not available. (b) Table 1: Results on the Extended YaleB database. Accuracy (%) Training time (s) Testing time (s) NSC 94.7 no need 1.41e-3 SVM e-5 CRC 97.0 no need 1.92e-3 SRC 96.5 no need 2.16e-2 DLSI(C) e-2 FDDL , LC-KSVD e-4 MMDL DPL e-4 Table 2: Results on the AR database. Accuracy (%) Training time (s) Testing time (s) NSC 92.0 no need 3.29e-3 SVM e-5 CRC 98.0 no need 5.08e-3 SRC 97.5 no need 3.42e-2 DLSI(C) , FDDL , LC-KSVD , e-4 MMDL DPL e-4 Extended YaleB database The recognition accuracies and training/testing time by different algorithms on the Extended YaleB database are summarized in Table 1. The proposed DPL algorithm achieves the best accuracy, which is slightly higher than MMDL, DLSI(C), LC-KSVD and FDDL. However, DPL has obvious advantage in efficiency over the other DL algorithms. AR database The recognition accuracies and running time on the AR database are shown in Table 2. DPL achieves the best results among all the competing algorithms. Compared with the experiment on Extended YaleB, in this experiment there are more training samples and the feature dimension is higher, and DPL s advantage of efficiency is much more obvious. In training, it is more than 159 times faster than DLSI and LC-KSVD, and 5,460 times faster than FDDL. 3.4 Object recognition In this section we test DPL on object categorization by using the Caltech101 database [25]. The Caltech101 database [25] includes 9,144 images from 102 classes (101 common object classes and a bacground class). The number of samples in each category varies from 31 to 800. Following the experimental settings in [5, 27], 30 samples per category are used for training and the rest are 7

8 Table 3: Recognition accuracy(%) & running time(s) on the Caltech101 database. Accuracy Training time Testing time NSC 70.1 no need 1.79e-2 SVM e-4 CRC 68.2 no need 1.38e-2 SRC 70.7 no need 1.09 DLSI(C) , FDDL , LC-KSVD , e-3 DPL e-3 used for testing. We use the standard bag-of-words (BOW) + spatial pyramid matching (SPM) framewor [27] for feature extraction. Dense SIFT descriptors are extracted on three grids of sizes 1 1, 2 2, and 4 4 to calculate the SPM features. For a fair comparison with [5], we use the vector quantization based coding method to extract the mid-level features and use the standard max pooling approach to build up the high dimension pooled features. Finally, the original 21,504 dimensional data is reduced to 3,000 dimension by PCA. The parameters τ and λ used in our algorithm are 0.05 and 1e-4, respectively. The experimental results are listed in Table 3. Again, DPL achieves the best performance. Though its classification accuracy is only slightly better than the DL methods, its advantage in terms of training/testing time is huge. 3.5 Action recognition Action recognition is an important yet very challenging tas and it has been attracting great research interests in recent years. We test our algorithm on the UCF 50 action database [26], which includes 50 categories of 6,680 human action videos from YouTube. We use the action ban features [28] and five-fold data splitting to evaluate our algorithm. For all the comparison methods, the feature dimension is reduced to 5,000 by PCA. The parameters τ and λ used in our algorithm are both The results by different methods are reported in Table 4. Our DPL algorithm achieves much higher accuracy than its competitors. FDDL has the second highest accuracy; however, it is 1,666 times slower than DPL in training and 83,317 times slower than DPL in testing. Table 4: Recognition accuracy(%) & running time(s) on the UCF50 action database Methods Accuracy Training time Testing time NSC 51.8 no need 6.11e-2 SVM e-4 CRC 60.3 no need 6.76e-2 SRC 59.6 no need 8.92 DLSI(C) , FDDL , LC-KSVD , DPL e-3 4 Conclusion We proposed a novel projective dictionary pair learning (DPL) model for pattern classification tass. Different from conventional dictionary learning (DL) methods, which learn a single synthesis dictionary, DPL learns jointly a synthesis dictionary and an analysis dictionary. Such a pair of dictionaries wor together to perform representation and discrimination simultaneously. Compared with previous DL methods, DPL employs projective coding, which largely reduces the computational burden in learning and testing. Performance evaluation was conducted on publically accessible visual classification datasets. DPL exhibits highly competitive classification accuracy with state-of-the-art DL methods, while it shows significantly higher efficiency, e.g., hundreds to thousands times faster than LC-KSVD and FDDL in training and testing. 8

9 References [1] Aharon, M., Elad, M., Brucstein, A.: K-svd: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. on Signal Processing, 54(11) (2006) [2] Wright, J., Yang, A.Y., Ganesh, A., Sastry, S.S., Ma, Y.: Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 31(2) (2009) [3] Rubinstein, R., Brucstein, A.M., Elad, M.: Dictionaries for sparse representation modeling. Proceedings of the IEEE 98(6) (2010) [4] Mairal, J., Bach, F., Ponce, J.: Tas-driven dictionary learning. IEEE Trans. Pattern Anal. Mach. Intelligence 34(4) (2012) [5] Jiang, Z., Lin, Z., Davis, L.: Label consistent -svd: learning a discriminative dictionary for recognition. IEEE Trans. on Pattern Anal. Mach. Intelligence 35(11) (2013) [6] Elad, M., Aharon, M.: Image denoising via sparse and redundant representations over learned dictionaries. IEEE Transactions on Image Processing 15(12) (2006) [7] Mairal, J., Bach, F., Ponce, J., Sapiro, G., Zisserman, A., et al.: Supervised dictionary learning. In: NIPS. (2008) [8] Ramirez, I., Sprechmann, P., Sapiro, G.: Classification and clustering via dictionary learning with structured incoherence and shared features. In: CVPR. (2010) [9] Yang, M., Zhang, L.,, Feng, X., Zhang, D.: Fisher discrimination dictionary learning for sparse representation. In: ICCV. (2011) [10] Wang, Z., Yang, J., Nasrabadi, N., Huang, T.: A max-margin perspective on sparse representation-based classification. In: ICCV. (2013) [11] Lee, H., Battle, A., Raina, R., Ng, A.Y.: Efficient sparse coding algorithms. In: NIPS. (2007) [12] Hale, E.T., Yin, W., Zhang, Y.: Fixed-point continuation for l 1-minimization: Methodology and convergence. SIAM Journal on Optimization 19(3) (2008) [13] Gregor, K., LeCun, Y.: Learning fast approximations of sparse coding. In: ICML. (2010) [14] Ranzato, M., Poultney, C., Chopra, S., Cun, Y.L.: Efficient learning of sparse representations with an energy-based model. In: NIPS. (2006) [15] Yunjin, C., Thomas, P., Bischof, H.: Learning l1-based analysis and synthesis sparsity priors using bilevel optimization. NIPS worshop (2012) [16] Elad, M., Milanfar, P., Rubinstein, R.: Analysis versus synthesis in signal priors. Inverse problems 23(3) (2007) 947 [17] Sprechmann, P., Litman, R., Yaar, T.B., Bronstein, A., Sapiro, G.: Efficient supervised sparse analysis and synthesis operators. In: NIPS. (2013) [18] Feng, Z., Yang, M., Zhang, L., Liu, Y., Zhang, D.: Joint discriminative dimensionality reduction and dictionary learning for face recognition. Pattern Recognition 46(8) (2013) [19] Soltanolotabi, M., Elhamifar, E., Candes, E.: Robust subspace clustering. arxiv preprint arxiv: (2013) [20] Coates, A., Ng, A.Y.: The importance of encoding versus training with sparse coding and vector quantization. In: ICML. (2011) [21] Zhang, L., Yang, M., Feng, X.: Sparse representation or collaborative representation: Which helps face recognition? In: ICCV. (2011) [22] Georghiades, A., Belhumeur, P., Kriegman, D.: From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Patt. Anal. Mach. Intel. 23(6) (2001) [23] Gorsi, J., Pfeuffer, F., Klamroth, K.: Biconvex sets and optimization with biconvex functions: a survey and extensions. Mathematical Methods of Operations Research 66(3) (2007) [24] Martinez, A., Benavente., R.: The ar face database. CVC Technical Report (1998) [25] Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. Computer Vision and Image Understanding 106(1) (2007) [26] Reddy, K.K., Shah, M.: Recognizing 50 human action categories of web videos. Machine Vision and Applications 24(5) (2013) [27] Lazebni, S., Schmid, C., Ponce, J.: Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: CVPR. (2006) [28] Sadanand, S., Corso, J.J.: Action ban: A high-level representation of activity in video. In: CVPR. (2012) 9

Virtual Training Samples and CRC based Test Sample Reconstruction and Face Recognition Experiments Wei HUANG and Li-ming MIAO

Virtual Training Samples and CRC based Test Sample Reconstruction and Face Recognition Experiments Wei HUANG and Li-ming MIAO 7 nd International Conference on Computational Modeling, Simulation and Applied Mathematics (CMSAM 7) ISBN: 978--6595-499-8 Virtual raining Samples and CRC based est Sample Reconstruction and Face Recognition

More information

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma

Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma Robust Face Recognition via Sparse Representation Authors: John Wright, Allen Y. Yang, Arvind Ganesh, S. Shankar Sastry, and Yi Ma Presented by Hu Han Jan. 30 2014 For CSE 902 by Prof. Anil K. Jain: Selected

More information

Discriminative Analysis Dictionary Learning

Discriminative Analysis Dictionary Learning Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence (AAAI-16 Discriative Analysis Dictionary Learning Jun Guo 1, Yanqing Guo 1, Xiangwei Kong 1, Man Zhang, and Ran He,3 1 School of

More information

Discriminative sparse model and dictionary learning for object category recognition

Discriminative sparse model and dictionary learning for object category recognition Discriative sparse model and dictionary learning for object category recognition Xiao Deng and Donghui Wang Institute of Artificial Intelligence, Zhejiang University Hangzhou, China, 31007 {yellowxiao,dhwang}@zju.edu.cn

More information

Joint Discriminative Dimensionality Reduction and Dictionary Learning for Face Recognition

Joint Discriminative Dimensionality Reduction and Dictionary Learning for Face Recognition Joint Discriminative Dimensionality Reduction and Dictionary Learning for Face Recognition Zhizhao Feng *, Meng Yang *, Lei Zhang 1, Yan Liu and David Zhang Dept. of Computing, The Hong Kong Polytechnic

More information

IMAGE SUPER-RESOLUTION BASED ON DICTIONARY LEARNING AND ANCHORED NEIGHBORHOOD REGRESSION WITH MUTUAL INCOHERENCE

IMAGE SUPER-RESOLUTION BASED ON DICTIONARY LEARNING AND ANCHORED NEIGHBORHOOD REGRESSION WITH MUTUAL INCOHERENCE IMAGE SUPER-RESOLUTION BASED ON DICTIONARY LEARNING AND ANCHORED NEIGHBORHOOD REGRESSION WITH MUTUAL INCOHERENCE Yulun Zhang 1, Kaiyu Gu 2, Yongbing Zhang 1, Jian Zhang 3, and Qionghai Dai 1,4 1 Shenzhen

More information

Face Recognition via Sparse Representation

Face Recognition via Sparse Representation Face Recognition via Sparse Representation John Wright, Allen Y. Yang, Arvind, S. Shankar Sastry and Yi Ma IEEE Trans. PAMI, March 2008 Research About Face Face Detection Face Alignment Face Recognition

More information

Robust Face Recognition via Sparse Representation

Robust Face Recognition via Sparse Representation Robust Face Recognition via Sparse Representation Panqu Wang Department of Electrical and Computer Engineering University of California, San Diego La Jolla, CA 92092 pawang@ucsd.edu Can Xu Department of

More information

IMA Preprint Series # 2281

IMA Preprint Series # 2281 DICTIONARY LEARNING AND SPARSE CODING FOR UNSUPERVISED CLUSTERING By Pablo Sprechmann and Guillermo Sapiro IMA Preprint Series # 2281 ( September 2009 ) INSTITUTE FOR MATHEMATICS AND ITS APPLICATIONS UNIVERSITY

More information

The Role of Dictionary Learning on Sparse Representation-Based Classification

The Role of Dictionary Learning on Sparse Representation-Based Classification The Role of Dictionary Learning on Sparse Representation-Based Classification Soheil Shafiee, Farhad Kamangar and Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington,

More information

Aggregating Descriptors with Local Gaussian Metrics

Aggregating Descriptors with Local Gaussian Metrics Aggregating Descriptors with Local Gaussian Metrics Hideki Nakayama Grad. School of Information Science and Technology The University of Tokyo Tokyo, JAPAN nakayama@ci.i.u-tokyo.ac.jp Abstract Recently,

More information

An efficient face recognition algorithm based on multi-kernel regularization learning

An efficient face recognition algorithm based on multi-kernel regularization learning Acta Technica 61, No. 4A/2016, 75 84 c 2017 Institute of Thermomechanics CAS, v.v.i. An efficient face recognition algorithm based on multi-kernel regularization learning Bi Rongrong 1 Abstract. A novel

More information

String distance for automatic image classification

String distance for automatic image classification String distance for automatic image classification Nguyen Hong Thinh*, Le Vu Ha*, Barat Cecile** and Ducottet Christophe** *University of Engineering and Technology, Vietnam National University of HaNoi,

More information

REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY. Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin

REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY. Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin REJECTION-BASED CLASSIFICATION FOR ACTION RECOGNITION USING A SPATIO-TEMPORAL DICTIONARY Stefen Chan Wai Tim, Michele Rombaut, Denis Pellerin Univ. Grenoble Alpes, GIPSA-Lab, F-38000 Grenoble, France ABSTRACT

More information

Regularized Robust Coding for Face Recognition

Regularized Robust Coding for Face Recognition Regularized Robust Coding for Face Recognition Meng Yang a, Student Member, IEEE, Lei Zhang a,, Member, IEEE Jian Yang b, Member, IEEE, and David Zhang a, Fellow, IEEE a Dept. of Computing, The Hong Kong

More information

arxiv: v1 [cs.lg] 20 Dec 2013

arxiv: v1 [cs.lg] 20 Dec 2013 Unsupervised Feature Learning by Deep Sparse Coding Yunlong He Koray Kavukcuoglu Yun Wang Arthur Szlam Yanjun Qi arxiv:1312.5783v1 [cs.lg] 20 Dec 2013 Abstract In this paper, we propose a new unsupervised

More information

LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA

LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES. Qiang Qiu and Guillermo Sapiro. Duke University, Durham, NC 27708, USA LEARNING COMPRESSED IMAGE CLASSIFICATION FEATURES Qiang Qiu and Guillermo Sapiro Duke University, Durham, NC 2778, USA ABSTRACT Learning a transformation-based dimension reduction, thereby compressive,

More information

Effectiveness of Sparse Features: An Application of Sparse PCA

Effectiveness of Sparse Features: An Application of Sparse PCA 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling

Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling Single Image Interpolation via Adaptive Non-Local Sparsity-Based Modeling Yaniv Romano The Electrical Engineering Department Matan Protter The Computer Science Department Michael Elad The Computer Science

More information

Image Restoration and Background Separation Using Sparse Representation Framework

Image Restoration and Background Separation Using Sparse Representation Framework Image Restoration and Background Separation Using Sparse Representation Framework Liu, Shikun Abstract In this paper, we introduce patch-based PCA denoising and k-svd dictionary learning method for the

More information

Multi-Class Image Classification: Sparsity Does It Better

Multi-Class Image Classification: Sparsity Does It Better Multi-Class Image Classification: Sparsity Does It Better Sean Ryan Fanello 1,2, Nicoletta Noceti 2, Giorgio Metta 1 and Francesca Odone 2 1 Department of Robotics, Brain and Cognitive Sciences, Istituto

More information

ANALYSIS SPARSE CODING MODELS FOR IMAGE-BASED CLASSIFICATION. Sumit Shekhar, Vishal M. Patel and Rama Chellappa

ANALYSIS SPARSE CODING MODELS FOR IMAGE-BASED CLASSIFICATION. Sumit Shekhar, Vishal M. Patel and Rama Chellappa ANALYSIS SPARSE CODING MODELS FOR IMAGE-BASED CLASSIFICATION Sumit Shekhar, Vishal M. Patel and Rama Chellappa Center for Automation Research, University of Maryland, College Park, MD 20742 {sshekha, pvishalm,

More information

Locality-Constrained Low-Rank Coding for Image Classification

Locality-Constrained Low-Rank Coding for Image Classification Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence Locality-Constrained Low-Rank Coding for Image Classification Ziheng Jiang, Ping Guo, Lihong Peng School of Computer Science

More information

Sparse Coding and Dictionary Learning for Image Analysis

Sparse Coding and Dictionary Learning for Image Analysis Sparse Coding and Dictionary Learning for Image Analysis Part IV: Recent Advances in Computer Vision and New Models Francis Bach, Julien Mairal, Jean Ponce and Guillermo Sapiro CVPR 10 tutorial, San Francisco,

More information

Discriminative Representative Selection via Structure Sparsity

Discriminative Representative Selection via Structure Sparsity 204 22nd International Conference on Pattern Recognition Discriative Representative Selection via Structure Sparsity Baoxing Wang, Qiyue Yin, Shu Wu, Liang Wang and Guiquan Liu School of Software and Engineering,

More information

Generalized Lasso based Approximation of Sparse Coding for Visual Recognition

Generalized Lasso based Approximation of Sparse Coding for Visual Recognition Generalized Lasso based Approximation of Sparse Coding for Visual Recognition Nobuyuki Morioka The University of New South Wales & NICTA Sydney, Australia nmorioka@cse.unsw.edu.au Shin ichi Satoh National

More information

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition

Tensor Decomposition of Dense SIFT Descriptors in Object Recognition Tensor Decomposition of Dense SIFT Descriptors in Object Recognition Tan Vo 1 and Dat Tran 1 and Wanli Ma 1 1- Faculty of Education, Science, Technology and Mathematics University of Canberra, Australia

More information

Sparse Variation Dictionary Learning for Face Recognition with A Single Training Sample Per Person

Sparse Variation Dictionary Learning for Face Recognition with A Single Training Sample Per Person Sparse Variation Dictionary Learning for Face Recognition with A Single Training Sample Per Person Meng Yang, Luc Van Gool ETH Zurich Switzerland {yang,vangool}@vision.ee.ethz.ch Lei Zhang The Hong Kong

More information

Multiple Kernel Learning for Emotion Recognition in the Wild

Multiple Kernel Learning for Emotion Recognition in the Wild Multiple Kernel Learning for Emotion Recognition in the Wild Karan Sikka, Karmen Dykstra, Suchitra Sathyanarayana, Gwen Littlewort and Marian S. Bartlett Machine Perception Laboratory UCSD EmotiW Challenge,

More information

Multi-Kernel Low-Rank Dictionary Pair Learning for Multiple Features Based Image Classification

Multi-Kernel Low-Rank Dictionary Pair Learning for Multiple Features Based Image Classification Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) Multi-Kernel Low-Rank Dictionary Pair Learning for Multiple Features Based Image Classification Xiaoke Zhu, 1,4 Xiao-Yuan

More information

SUPPLEMENTARY MATERIAL

SUPPLEMENTARY MATERIAL SUPPLEMENTARY MATERIAL Zhiyuan Zha 1,3, Xin Liu 2, Ziheng Zhou 2, Xiaohua Huang 2, Jingang Shi 2, Zhenhong Shang 3, Lan Tang 1, Yechao Bai 1, Qiong Wang 1, Xinggan Zhang 1 1 School of Electronic Science

More information

Supervised Translation-Invariant Sparse Coding

Supervised Translation-Invariant Sparse Coding Supervised Translation-Invariant Sparse Coding Jianchao Yang,KaiYu, Thomas Huang Beckman Institute, University of Illinois at Urbana-Champaign NEC Laboratories America, Inc., Cupertino, California {jyang29,

More information

Linear Discriminant Analysis for 3D Face Recognition System

Linear Discriminant Analysis for 3D Face Recognition System Linear Discriminant Analysis for 3D Face Recognition System 3.1 Introduction Face recognition and verification have been at the top of the research agenda of the computer vision community in recent times.

More information

Bilevel Sparse Coding

Bilevel Sparse Coding Adobe Research 345 Park Ave, San Jose, CA Mar 15, 2013 Outline 1 2 The learning model The learning algorithm 3 4 Sparse Modeling Many types of sensory data, e.g., images and audio, are in high-dimensional

More information

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING

IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING IMAGE DENOISING USING NL-MEANS VIA SMOOTH PATCH ORDERING Idan Ram, Michael Elad and Israel Cohen Department of Electrical Engineering Department of Computer Science Technion - Israel Institute of Technology

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

Learning based face hallucination techniques: A survey

Learning based face hallucination techniques: A survey Vol. 3 (2014-15) pp. 37-45. : A survey Premitha Premnath K Department of Computer Science & Engineering Vidya Academy of Science & Technology Thrissur - 680501, Kerala, India (email: premithakpnath@gmail.com)

More information

Learning-based Methods in Vision

Learning-based Methods in Vision Learning-based Methods in Vision 16-824 Sparsity and Deep Learning Motivation Multitude of hand-designed features currently in use in vision - SIFT, HoG, LBP, MSER, etc. Even the best approaches, just

More information

arxiv: v3 [cs.cv] 3 Oct 2012

arxiv: v3 [cs.cv] 3 Oct 2012 Combined Descriptors in Spatial Pyramid Domain for Image Classification Junlin Hu and Ping Guo arxiv:1210.0386v3 [cs.cv] 3 Oct 2012 Image Processing and Pattern Recognition Laboratory Beijing Normal University,

More information

Beyond Bags of Features

Beyond Bags of Features : for Recognizing Natural Scene Categories Matching and Modeling Seminar Instructed by Prof. Haim J. Wolfson School of Computer Science Tel Aviv University December 9 th, 2015

More information

Efficient Implementation of the K-SVD Algorithm and the Batch-OMP Method

Efficient Implementation of the K-SVD Algorithm and the Batch-OMP Method Efficient Implementation of the K-SVD Algorithm and the Batch-OMP Method Ron Rubinstein, Michael Zibulevsky and Michael Elad Abstract The K-SVD algorithm is a highly effective method of training overcomplete

More information

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features

Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Supplementary material: Strengthening the Effectiveness of Pedestrian Detection with Spatially Pooled Features Sakrapee Paisitkriangkrai, Chunhua Shen, Anton van den Hengel The University of Adelaide,

More information

Hierarchical Matching Pursuit for Image Classification: Architecture and Fast Algorithms

Hierarchical Matching Pursuit for Image Classification: Architecture and Fast Algorithms Hierarchical Matching Pursuit for Image Classification: Architecture and Fast Algorithms Liefeng Bo University of Washington Seattle WA 98195, USA Xiaofeng Ren ISTC-Pervasive Computing Intel Labs Seattle

More information

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization

Image Deblurring Using Adaptive Sparse Domain Selection and Adaptive Regularization Volume 3, No. 3, May-June 2012 International Journal of Advanced Research in Computer Science RESEARCH PAPER Available Online at www.ijarcs.info ISSN No. 0976-5697 Image Deblurring Using Adaptive Sparse

More information

Sparse Models in Image Understanding And Computer Vision

Sparse Models in Image Understanding And Computer Vision Sparse Models in Image Understanding And Computer Vision Jayaraman J. Thiagarajan Arizona State University Collaborators Prof. Andreas Spanias Karthikeyan Natesan Ramamurthy Sparsity Sparsity of a vector

More information

An Iteratively Reweighted Least Square Implementation for Face Recognition

An Iteratively Reweighted Least Square Implementation for Face Recognition Vol. 6: 26-32 THE UNIVERSITY OF CENTRAL FLORIDA Published May 15, 2012 An Iteratively Reweighted Least Square Implementation for Face Recognition By: Jie Liang Faculty Mentor: Dr. Xin Li ABSTRACT: We propose,

More information

Facial Expression Recognition Using Non-negative Matrix Factorization

Facial Expression Recognition Using Non-negative Matrix Factorization Facial Expression Recognition Using Non-negative Matrix Factorization Symeon Nikitidis, Anastasios Tefas and Ioannis Pitas Artificial Intelligence & Information Analysis Lab Department of Informatics Aristotle,

More information

In Defense of Sparsity Based Face Recognition

In Defense of Sparsity Based Face Recognition 013 IEEE Conference on Computer Vision and Pattern Recognition In Defense of Sparsity Based Face Recognition Weihong Deng, Jiani Hu, Jun Guo Beijing University of Posts and Telecommunications, Beijing,

More information

Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification?

Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification? Supplementary material for the paper Are Sparse Representations Really Relevant for Image Classification? Roberto Rigamonti, Matthew A. Brown, Vincent Lepetit CVLab, EPFL Lausanne, Switzerland firstname.lastname@epfl.ch

More information

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY

More information

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation

A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation , pp.162-167 http://dx.doi.org/10.14257/astl.2016.138.33 A Novel Image Super-resolution Reconstruction Algorithm based on Modified Sparse Representation Liqiang Hu, Chaofeng He Shijiazhuang Tiedao University,

More information

Action Recognition By Learnt Class-Specific Overcomplete Dictionaries

Action Recognition By Learnt Class-Specific Overcomplete Dictionaries Action Recognition By Learnt Class-Specific Overcomplete Dictionaries Tanaya Guha Electrical and Computer Engineering University of British Columbia Vancouver, Canada Email: tanaya@ece.ubc.ca Rabab K.

More information

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising

An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising An Optimized Pixel-Wise Weighting Approach For Patch-Based Image Denoising Dr. B. R.VIKRAM M.E.,Ph.D.,MIEEE.,LMISTE, Principal of Vijay Rural Engineering College, NIZAMABAD ( Dt.) G. Chaitanya M.Tech,

More information

ImageCLEF 2011

ImageCLEF 2011 SZTAKI @ ImageCLEF 2011 Bálint Daróczy joint work with András Benczúr, Róbert Pethes Data Mining and Web Search Group Computer and Automation Research Institute Hungarian Academy of Sciences Training/test

More information

Experiments of Image Retrieval Using Weak Attributes

Experiments of Image Retrieval Using Weak Attributes Columbia University Computer Science Department Technical Report # CUCS 005-12 (2012) Experiments of Image Retrieval Using Weak Attributes Felix X. Yu, Rongrong Ji, Ming-Hen Tsai, Guangnan Ye, Shih-Fu

More information

on learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015

on learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015 on learned visual embedding patrick pérez Allegro Workshop Inria Rhônes-Alpes 22 July 2015 Vector visual representation Fixed-size image representation High-dim (100 100,000) Generic, unsupervised: BoW,

More information

Synthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University

Synthesis and Analysis Sparse Representation Models for Image Restoration. Shuhang Gu 顾舒航. Dept. of Computing The Hong Kong Polytechnic University Synthesis and Analysis Sparse Representation Models for Image Restoration Shuhang Gu 顾舒航 Dept. of Computing The Hong Kong Polytechnic University Outline Sparse representation models for image modeling

More information

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method Journal of Information Hiding and Multimedia Signal Processing c 2016 ISSN 2073-4212 Ubiquitous International Volume 7, Number 5, September 2016 Face Recognition ased on LDA and Improved Pairwise-Constrained

More information

Spatial Locality-Aware Sparse Coding and Dictionary Learning

Spatial Locality-Aware Sparse Coding and Dictionary Learning JMLR: Workshop and Conference Proceedings 25:491 505, 2012 Asian Conference on Machine Learning Spatial Locality-Aware Sparse Coding and Dictionary Learning Jiang Wang 2145 Sheridan Road, Evanston IL 60208

More information

Sparse coding for image classification

Sparse coding for image classification Sparse coding for image classification Columbia University Electrical Engineering: Kun Rong(kr2496@columbia.edu) Yongzhou Xiang(yx2211@columbia.edu) Yin Cui(yc2776@columbia.edu) Outline Background Introduction

More information

Learning Convolutional Feature Hierarchies for Visual Recognition

Learning Convolutional Feature Hierarchies for Visual Recognition Learning Convolutional Feature Hierarchies for Visual Recognition Koray Kavukcuoglu, Pierre Sermanet, Y-Lan Boureau, Karol Gregor, Michael Mathieu, Yann LeCun Computer Science Department Courant Institute

More information

Part-based and local feature models for generic object recognition

Part-based and local feature models for generic object recognition Part-based and local feature models for generic object recognition May 28 th, 2015 Yong Jae Lee UC Davis Announcements PS2 grades up on SmartSite PS2 stats: Mean: 80.15 Standard Dev: 22.77 Vote on piazza

More information

Multimodal Dictionary Learning and Joint Sparse Representation for HEp-2 Cell Classification

Multimodal Dictionary Learning and Joint Sparse Representation for HEp-2 Cell Classification Multimodal Dictionary Learning and Joint Sparse Representation for HEp-2 Cell Classification Ali Taalimi 1, Shahab Ensafi 2,3, Hairong Qi 1, Shijian Lu 2, Ashraf A. Kassim 3, and Chew Lim Tan 4 1 University

More information

Exploring Bag of Words Architectures in the Facial Expression Domain

Exploring Bag of Words Architectures in the Facial Expression Domain Exploring Bag of Words Architectures in the Facial Expression Domain Karan Sikka, Tingfan Wu, Josh Susskind, and Marian Bartlett Machine Perception Laboratory, University of California San Diego {ksikka,ting,josh,marni}@mplab.ucsd.edu

More information

A Representative Sample Selection Approach for SRC

A Representative Sample Selection Approach for SRC DEIM Forum 2011 E9-1 AliceChen, NTT, 239-0847 1-1 School of Computing Science, Simon Fraser University 8888 University Drive, Burnaby BC, V5A 1S6 Canada E-mail: {alice.chen,eda.takeharu,katafuchi.norifumi,kataoka.ryoji}@lab.ntt.co.jp

More information

Modeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations

Modeling Visual Cortex V4 in Naturalistic Conditions with Invari. Representations Modeling Visual Cortex V4 in Naturalistic Conditions with Invariant and Sparse Image Representations Bin Yu Departments of Statistics and EECS University of California at Berkeley Rutgers University, May

More information

Sparse Coding for Learning Interpretable Spatio-Temporal Primitives

Sparse Coding for Learning Interpretable Spatio-Temporal Primitives Sparse Coding for Learning Interpretable Spatio-Temporal Primitives Taehwan Kim TTI Chicago taehwan@ttic.edu Gregory Shakhnarovich TTI Chicago gregory@ttic.edu Raquel Urtasun TTI Chicago rurtasun@ttic.edu

More information

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision

Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Supplementary Material : Partial Sum Minimization of Singular Values in RPCA for Low-Level Vision Due to space limitation in the main paper, we present additional experimental results in this supplementary

More information

arxiv: v1 [cs.cv] 1 Aug 2013

arxiv: v1 [cs.cv] 1 Aug 2013 Domain-invariant Face Recognition using Learned Low-rank Transformation arxiv:1308.0275v1 [cs.cv] 1 Aug 2013 Abstract Qiang Qiu Duke University Durham, NC, 27708 qiang.qiu@duke.edu Guillermo Sapiro Duke

More information

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li

Learning to Match. Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li Learning to Match Jun Xu, Zhengdong Lu, Tianqi Chen, Hang Li 1. Introduction The main tasks in many applications can be formalized as matching between heterogeneous objects, including search, recommendation,

More information

Face Recognition by Combining Kernel Associative Memory and Gabor Transforms

Face Recognition by Combining Kernel Associative Memory and Gabor Transforms Face Recognition by Combining Kernel Associative Memory and Gabor Transforms Author Zhang, Bai-ling, Leung, Clement, Gao, Yongsheng Published 2006 Conference Title ICPR2006: 18th International Conference

More information

Geodesic Flow Kernel for Unsupervised Domain Adaptation

Geodesic Flow Kernel for Unsupervised Domain Adaptation Geodesic Flow Kernel for Unsupervised Domain Adaptation Boqing Gong University of Southern California Joint work with Yuan Shi, Fei Sha, and Kristen Grauman 1 Motivation TRAIN TEST Mismatch between different

More information

Robust Multimodal Dictionary Learning

Robust Multimodal Dictionary Learning Robust Multimodal Dictionary Learning Tian Cao 1, Vladimir Jojic 1, Shannon Modla 3, Debbie Powell 3, Kirk Czymmek 4, and Marc Niethammer 1,2 1 University of North Carolina at Chapel Hill, NC 2 Biomedical

More information

Beyond Spatial Pyramids: Receptive Field Learning for Pooled Image Features

Beyond Spatial Pyramids: Receptive Field Learning for Pooled Image Features Beyond Spatial Pyramids: Receptive Field Learning for Pooled Image Features Yangqing Jia UC Berkeley EECS jiayq@berkeley.edu Chang Huang NEC Labs America chuang@sv.nec-labs.com Abstract We examine the

More information

Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition

Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition Salient Region Detection and Segmentation in Images using Dynamic Mode Decomposition Sikha O K 1, Sachin Kumar S 2, K P Soman 2 1 Department of Computer Science 2 Centre for Computational Engineering and

More information

Generalized Domain-Adaptive Dictionaries

Generalized Domain-Adaptive Dictionaries 2013 IEEE Conference on Computer Vision and Pattern Recognition Generalized Domain-Adaptive Dictionaries Sumit Shekhar Vishal M. Patel Hien V. Nguyen Rama Chellappa University of Maryland, College Park,

More information

FEATURE learning is a critical step for almost all

FEATURE learning is a critical step for almost all 1080 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 27, NO. 5, MAY 2016 Integrated Low-Rank-Based Discriminative Feature Learning for Recognition Pan Zhou, Zhouchen Lin, Senior Member,

More information

AUTOMATIC data summarization, which attempts to

AUTOMATIC data summarization, which attempts to 748 JOURNAL OF SOFTWARE, VOL. 9, NO. 3, MARCH 2014 Sparse Affinity Propagation for Image Analysis Xue Zhang, Jian Cheng Lv Machine Intelligence Laboratory, College of Computer Science, Sichuan University,

More information

Pattern Recognition Letters Authorship Confirmation

Pattern Recognition Letters Authorship Confirmation Pattern Recognition Letters Authorship Confirmation Please save a copy of this file, complete and upload as the Confirmation of Authorship file. As corresponding author I, Mohammadreaza Babaee, hereby

More information

Semi-Supervised PCA-based Face Recognition Using Self-Training

Semi-Supervised PCA-based Face Recognition Using Self-Training Semi-Supervised PCA-based Face Recognition Using Self-Training Fabio Roli and Gian Luca Marcialis Dept. of Electrical and Electronic Engineering, University of Cagliari Piazza d Armi, 09123 Cagliari, Italy

More information

An Improved Approach For Mixed Noise Removal In Color Images

An Improved Approach For Mixed Noise Removal In Color Images An Improved Approach For Mixed Noise Removal In Color Images Ancy Mariam Thomas 1, Dr. Deepa J 2, Rijo Sam 3 1P.G. student, College of Engineering, Chengannur, Kerala, India. 2Associate Professor, Electronics

More information

Three things everyone should know to improve object retrieval. Relja Arandjelović and Andrew Zisserman (CVPR 2012)

Three things everyone should know to improve object retrieval. Relja Arandjelović and Andrew Zisserman (CVPR 2012) Three things everyone should know to improve object retrieval Relja Arandjelović and Andrew Zisserman (CVPR 2012) University of Oxford 2 nd April 2012 Large scale object retrieval Find all instances of

More information

Improving Recognition through Object Sub-categorization

Improving Recognition through Object Sub-categorization Improving Recognition through Object Sub-categorization Al Mansur and Yoshinori Kuno Graduate School of Science and Engineering, Saitama University, 255 Shimo-Okubo, Sakura-ku, Saitama-shi, Saitama 338-8570,

More information

Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels

Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels Action Recognition in Video by Sparse Representation on Covariance Manifolds of Silhouette Tunnels Kai Guo, Prakash Ishwar, and Janusz Konrad Department of Electrical & Computer Engineering Motivation

More information

On Compact Codes for Spatially Pooled Features

On Compact Codes for Spatially Pooled Features Yangqing Jia Oriol Vinyals Trevor Darrell UC Berkeley EECS, Berkeley, CA 97 USA jiayq@eecs.berkeley.edu vinyals@eecs.berkeley.edu trevor@eecs.berkeley.edu Abstract Feature encoding with an overcomplete

More information

Sketchable Histograms of Oriented Gradients for Object Detection

Sketchable Histograms of Oriented Gradients for Object Detection Sketchable Histograms of Oriented Gradients for Object Detection No Author Given No Institute Given Abstract. In this paper we investigate a new representation approach for visual object recognition. The

More information

Image-Based Face Recognition using Global Features

Image-Based Face Recognition using Global Features Image-Based Face Recognition using Global Features Xiaoyin xu Research Centre for Integrated Microsystems Electrical and Computer Engineering University of Windsor Supervisors: Dr. Ahmadi May 13, 2005

More information

Novel Lossy Compression Algorithms with Stacked Autoencoders

Novel Lossy Compression Algorithms with Stacked Autoencoders Novel Lossy Compression Algorithms with Stacked Autoencoders Anand Atreya and Daniel O Shea {aatreya, djoshea}@stanford.edu 11 December 2009 1. Introduction 1.1. Lossy compression Lossy compression is

More information

Learning Algorithms for Medical Image Analysis. Matteo Santoro slipguru

Learning Algorithms for Medical Image Analysis. Matteo Santoro slipguru Learning Algorithms for Medical Image Analysis Matteo Santoro slipguru santoro@disi.unige.it June 8, 2010 Outline 1. learning-based strategies for quantitative image analysis 2. automatic annotation of

More information

Heat Kernel Based Local Binary Pattern for Face Representation

Heat Kernel Based Local Binary Pattern for Face Representation JOURNAL OF LATEX CLASS FILES 1 Heat Kernel Based Local Binary Pattern for Face Representation Xi Li, Weiming Hu, Zhongfei Zhang, Hanzi Wang Abstract Face classification has recently become a very hot research

More information

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro

Learning Low-rank Transformations: Algorithms and Applications. Qiang Qiu Guillermo Sapiro Learning Low-rank Transformations: Algorithms and Applications Qiang Qiu Guillermo Sapiro Motivation Outline Low-rank transform - algorithms and theories Applications Subspace clustering Classification

More information

Object Classification Problem

Object Classification Problem HIERARCHICAL OBJECT CATEGORIZATION" Gregory Griffin and Pietro Perona. Learning and Using Taxonomies For Fast Visual Categorization. CVPR 2008 Marcin Marszalek and Cordelia Schmid. Constructing Category

More information

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING

NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING NULL SPACE CLUSTERING WITH APPLICATIONS TO MOTION SEGMENTATION AND FACE CLUSTERING Pan Ji, Yiran Zhong, Hongdong Li, Mathieu Salzmann, Australian National University, Canberra NICTA, Canberra {pan.ji,hongdong.li}@anu.edu.au,mathieu.salzmann@nicta.com.au

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

Probabilistic Location Recognition using Reduced Feature Set

Probabilistic Location Recognition using Reduced Feature Set Probabilistic Location Recognition using Reduced Feature Set Fayin Li and Jana Košecá Department of Computer Science George Mason University, Fairfax, VA 3 Email: {fli,oseca}@cs.gmu.edu Abstract The localization

More information

SP-SVM: Large Margin Classifier for Data on Multiple Manifolds

SP-SVM: Large Margin Classifier for Data on Multiple Manifolds : Large Margin Classifier for Data on Multiple Manifolds Bin Shen, Bao-Di Liu, Qifan Wang, Yi Fang, Jan P. Allebach Department of Computer Science, Purdue University, West Lafayette, IN. 47907, USA College

More information

HIGH-dimensional data are commonly observed in various

HIGH-dimensional data are commonly observed in various 1 Simplex Representation for Subspace Clustering Jun Xu 1, Student Member, IEEE, Deyu Meng 2, Member, IEEE, Lei Zhang 1, Fellow, IEEE 1 Department of Computing, The Hong Kong Polytechnic University, Hong

More information

Action Recognition Using Super Sparse Coding Vector with Spatio-Temporal Awareness

Action Recognition Using Super Sparse Coding Vector with Spatio-Temporal Awareness Action Recognition Using Super Sparse Coding Vector with Spatio-Temporal Awareness Xiaodong Yang and YingLi Tian Department of Electrical Engineering City College, City University of New York Abstract.

More information

Combining Selective Search Segmentation and Random Forest for Image Classification

Combining Selective Search Segmentation and Random Forest for Image Classification Combining Selective Search Segmentation and Random Forest for Image Classification Gediminas Bertasius November 24, 2013 1 Problem Statement Random Forest algorithm have been successfully used in many

More information

A A A. Fig.1 image patch. Then the edge gradient magnitude is . (1)

A A A. Fig.1 image patch. Then the edge gradient magnitude is . (1) International Conference on Information Science and Computer Applications (ISCA 013) Two-Dimensional Barcode Image Super-Resolution Reconstruction Via Sparse Representation Gaosheng Yang 1,Ningzhong Liu

More information