IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1

Size: px
Start display at page:

Download "IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1"

Transcription

1 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1 Exploring Locally Adaptive Dimensionality Reduction for Hyperspectral Image Classification: A Maximum Margin Metric Learning Aspect Yanni Dong, Student Member, IEEE,BoDu, Senior Member, IEEE, Liangpei Zhang, Senior Member, IEEE, and Lefei Zhang, Member, IEEE Abstract The high-dimensional data space generated by hyperspectral sensors introduces challenges for the conventional data analysis techniques. Popular dimensionality reduction techniques usually assume a Gaussian distribution, which may not be in accordance with real life. Metric learning methods, which explore the global data structure of the labeled training samples, have proved to be very efficient in hyperspectral fields. However, we can go further by utilizing locally adaptive decision constraints for the labeled training samples per class to obtain an even better performance. In this paper, we present the locally adaptive dimensionality reduction metric learning (LADRml) method for hyperspectral image classification. The aims of the presented method are: 1) first, to utilize the limited training samples to reduce the dimensionality of data without a certain distribution hypothesis; and 2) second, to better handle data with complex distributions by the use of locally adaptive decision constraints, which can assess the similarity between a pair of samples based on the distance changes before and after metric learning. The experimental results obtained with a number of challenging hyperspectral image datasets demonstrate that the proposed LADRml algorithm outperforms the state-of-the-art dimensionality reduction and metric learning methods. Index Terms Dimensionality reduction, hyperspectral image classification, locally adaptive decision constraints, metric learning. I. INTRODUCTION HYPERSPECTRAL sensors, which can gather hyperspectral imagery with hundreds of spectral bands, have been widely used for discriminating the subtle differences in ground objects. The rich bands contain much helpful information for image classification and have driven the development of advanced image classification techniques [1] [3]. Hyperspectral image classification, which is aimed at determining a unique label for each pixel to generate a thematic land-cover map, is Manuscript received May 09, 2016; revised June 23, 2016; accepted June 29, This work was supported in part by the National Basic Research Program of China (973 Program) under Grant 2012CB719905, the National Natural Science Foundation of China under Grants , , , and , the Natural Science Foundation of Hubei Province under Grant 2014CFB193, and the Fundamental Research Funds for the Central Universities. (Corresponding Author: Liangpei Zhang.) Y. Dong and L. Zhang are with the State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing, Wuhan University , China ( dongyanni@whu.edu.cn; zlp62@whu.edu.cn). B. Du is with the School of Computer, Wuhan University, Wuhan , China ( gunspace@163.com). L. Zhang is with the Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong ( cslfzhang@comp.polyu.edu.hk). Color versions of one or more of the figures in this paper are available online at Digital Object Identifier /JSTARS one of the most common tasks in hyperspectral image analysis [4] [8]. However, several critical problems still remain. First, the huge number of spectral bands may result in the Hughes phenomenon (a decrease in the classification accuracy when increasing the number of spectral bands, due to the low ratio between the number of training samples and the spectral channels) [9], [10]. That is, the high-dimensional data space generated by the hyperspectral sensors creates a new challenge, which is the significant computational cost caused by the data complexity. Thus, hyperspectral image classification usually follows dimensionality reduction, which is aimed at reducing the dimensionality of the feature space to decrease the computational complexity and discard the redundant features. Dimensionality reduction technology allows for the separation of classes without sacrificing significant information [11]. Second, the training samples are limited, which may result in overfitting of the training data [12], [13]. Researchers have implied that high-dimensional data spaces are mostly empty, which indicates that the useful data exist primarily in a subspace. From this aspect, dimensionality reduction also needs to be considered [14]. In the past decades, a large number of dimensionality reduction methods have been proposed to transform data from a high dimension to a low one. In general, dimensionality reduction techniques can be categorized into unsupervised and supervised approaches. The most representative unsupervised dimensionality reduction methods are principal component analysis (PCA) [15], [16], independent component analysis (ICA) [17], and locality preserving projections (LPP) [18]. PCA maximizes the amount of data variance in the projected linear subspace, while ICA uses higher order statistics. LPP keeps the local geometric structure of the original feature space to reduce the dimension. The typical representative supervised approaches include Fisher s linear discriminant analysis (LDA) and its variants, e. g., local Fisher s discriminant analysis (LFDA) [11] and sparse discriminant analysis (SDA) [19]. LFDA preserves the underlying structure of the multimodal non-gaussian class distribution in the projection, while SDA regularizes the usual LDA loss function by adding an L1 constraint to the weights. Discriminative locality alignment (DLA) also belongs to the supervised approaches, and it involves selecting both intraclass and interclass neighbors for a local patch to enlarge the margin between different classes [20]. As to hyperspectral image classification, many different approaches have been proposed in recent years, e.g., traditional dimensionality reduction [2], [21], sparse representation IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See standards/publications/rights/index.html for more information.

2 2 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING Fig. 1. Illustration of exploiting the locally adaptive decision constraints. [22] [24], semisupervised learning (SSL) [25] [27], and transfer learning [28] [33]. In essence, these methods assess the similarity between spectral signatures. That is, the above methods can be boiled down to new insights into a metric. For example: 1) traditional dimensionality reduction seeks a lowdimensional representation from a high-dimensional space by assuming a Gaussian distribution to discriminate the betweenclass and within-class distances in a new feature space; 2) sparse representation, which involves studying the relationship between samples in the sparse feature space, can be compactly represented by a few coefficients in a certain dictionary; and 3) SSL considers both labeled and unlabeled samples to obtain a better dividing hyperplane which can measure the distances between samples. That is to say, the existing methods for solving high-dimensional problems can be summarized as learning a more stable and credible distance metric. In fact, metric learning methods have proved to be a more straightforward and effective way to obtain such a distance metric [34], [35]. A number of metric learning methods have been developed, such as the relevant component analysis (RCA) method [36], which is a simple and efficient algorithm proposed for learning a global linear transformation by exploiting only the equivalence to reduce the irrelevant variability of the data. The neighborhood component analysis (NCA) method defines the probability of selecting the same class instances as the neighbors for each instance. It also encourages instances from the same class to be close, which can maximize the stochastic variance of the leave-one-out k-nearest neighbor (KNN) score on the training samples [37]. The large margin nearest neighbor (LMNN) method [38] aims to find a distance metric such that the instances from different classes are effectively separated by a large margin within the neighborhood, where the margin is defined as the difference between the between-class and within-class distances. Furthermore, the information-theoretic metric learning (ITML) method was proposed to express the weakly supervised metric learning problem as a Bregman optimization problem. ITML can handle a variety of constraints and incorporate a prior on the distance function [39]. By metric learning, we can find a distance metric which can transform the high dimension to a low one to classify the images by effectively maximizing the between-class distance while minimizing the within-class distance. Due to these characteristics, metric learning algorithms can also be used for dimensionality reduction, and they have been used to solve a range of problems in hyperspectral image analysis, such as feature extraction [40], image segmentation [41], and target detection [42], [43]. Among them, our previous work [maximum margin metric learning (MMML)] [43] is very different from the method presented in this paper (see Section II for a summary of the differences between the two methods). However, the current metric learning methods still have some obstacles that need to be addressed, e.g., the RCA algorithm lacks negative (dissimilarity) constraints, which can be informative, and it cannot capture complex nonlinear relationships between data instances; the NCA algorithm cannot obtain the optimal value of the objective function if the initial point is not selected appropriately, and it has a relatively high computational complexity; and the LMNN method has problems when dealing with high-dimensional data and adjusting parameters [44], [45]. Moreover, the state-of-the-art metric learning-based algorithms aim to ensure that samples from the same class are closer to each other than those from different classes, and make decisions based on comparing their Mahalanobis distance d and a fixed threshold b. In other words, the metric learning methods only lead to an absolute decision rule with a fixed threshold b, and they need relative constraints between the between-class and within-class pairs. To deal with the problem of high-dimensional data and to reduce the computational burden, a mapping function can be adopted from the high-dimensional input space into a low-dimensional embedding. Furthermore, a smoothness constraint can be represented by a regularization term, while a cutting plane algorithm can be used to optimize the algorithm in a constant number of iterations by taking the form of a large number of pairwise constraints with similar or dissimilar labels. However, the existing metric learning methods usually look for a distance measure and make a decision based on a fixed threshold, which is insufficient and suboptimal. Moreover, a further problem in this situation is that the between-class and withinclass variations are complex, meaning that these methods may not learn effective metrics for data with complex distributions. The processing is also difficult, because of the complex data. Consequently, the classification results are not accurate enough [46] [49]. Aiming at both the above problems, how can we go further by elaborately constructing a more promising metric to obtain a better classification performance than the state-of-theart dimensionality reduction and metric learning methods? In this paper, the locally adaptive dimensionality reduction metric learning (LADRml) method is proposed for adaptively classifying hyperspectral images to further improve the classification accuracy. Locally adaptive decision constraints are applied to relax the fixed threshold, which allows us to make a decision by considering both the threshold b and the changes between the distances before and after metric learning. We can go further by utilizing the local information of the labeled training samples per class to obtain a better performance. The contributions of this paper can be summarized as follows. 1) For the first time, the proposed algorithm applies locally adaptive decision constraints to enhance the separability between different classes. It can guarantee the distance between different classes to the greatest degree by

3 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 3 Fig. 2. Classification OA of the proposed method with respect to parameter C for the three datasets. (a) AVIRIS Indian Pines dataset. (b) Washington DC Mall dataset. (c) ROSIS Pavia University dataset. Fig. 3. (a) False-color image of the AVIRIS Indian Pines scene. (b) Ground-truth map containing 16 mutually exclusive classes. (c) Training samples used in the experiment. TABLE I NUMBERS OF SAMPLES OF THE AVIRIS INDIAN PINES DATASET USED FOR THEEXPERIMENT considering the distance between a pair of samples based on a threshold and the changes between the distances before and after metric learning. 2) In addition, the proposed method can effectively and efficiently encode the discriminative information from limited training data by considering the locality of data distribution, which considers neighboring constraints and avoids adopting those conflicting constraints. 3) Thus, we combine a global metric and the locally adaptive decision constraints using a joint MMML model, for which the number of parameters is less than the other MMML methods. As a result, the proposed algorithm makes fewer assumptions of the data and has strong generalization ability for dimensionality reduction. The remainder of this paper is organized as follows. Section II reviews the conventional regularized distance metric learning theory and details the proposed method with a step-by-step formula derivation. The experimental results of three challenging real-world hyperspectral datasets are presented to conduct a comparison between the different algorithms in Section III, followed by the conclusions in Section IV. II. LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION METRIC LEARNING METHOD In this section, we first state the general problem and the important concept of metric learning. After this, the formulation and characteristics of the locally adaptive decision constraints are detailed. Finally, the LADRml method is proposed in the last section. In this section, we consider the following metric learning algorithm by taking a set of labeled training samples S = {(x 1,y 1 ), (x 2,y 2 ),...(x n,y n )} as the input, where [x 1,...,x n ] R L n. x i is the ith input data point and y i is its corresponding discrete label. L is the number of features and n denotes the number of samples. We then have the following two pairwise constraint sets: a set of similar constraints Λ and a set of dissimilar constraints Ω [50] [52]: Λ: (x i, x j ) Λ,x i, x j same class (1) Ω: (x i, x j ) Ω,x i, x j different class.

4 4 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING TABLE II INDIVIDUAL CLASS ACCURACIES,OAS, KAPPA STATISTIC VALUES, AND RUNNING TIMES (IN S) OF THE AVIRIS INDIAN PINES DATASET OBTAINED BY THE DIFFERENT CLASSIFICATION METHODS (dim = 60) Class Original MNF-SVM SDA DLA LDA LFDA RCA NCA LMNN ITML MMML LADR ml Alfalfa Corn-no till Corn-min till Corn Grass/pasture Grass/tree Grass/pasture-mowed Hay-windrowed Oats Soybeans-no till Soybeans-min till Soybeans-clean till Wheat Woods Bldg-grass-tree-drives Stone-steel towers OA 64.93± 67.45± 66.83± 72.74± 58.63± 71.52± 51.15± 68.58± 65.23± 63.30± 69.14± 75.80± Kappa 0.596± 0.630± 0.618± 0.690± 0.545± 0.677± 0.441± 0.641± 0.609± 0.580± 0.648± 0.723± Time Methods Fig. 4. Classification maps obtained from the AVIRIS Indian Pines dataset, along with the training set selected by each method for: (a) Original, (b) MNF-SVM, (c) SDA, (d) DLA, (e) LDA, (f) LFDA, (g) RCA, (h) NCA, (i) LMNN, (j) ITML, (k) MMML, and (l) LADRml.

5 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 5 With (1), we can rewrite the above formulation as d W (x i, x j ) b(x i, x j ) Λ d W (x i, x j ) b(x i, x j ) Ω. (5) Fig. 5. Classification OA with regard to reduced dimensionality in the AVIRIS Indian Pines dataset. The goal of metric learning is to learn a positive semidefinite (PSD) matrix M, which specifies the Mahalanobis distance d M (x i, x j ), such that the distance between x i and x j can be computed from: d M (x i, x j )= (x i x j ) T M(x i x j ) = M, (x i x j )(x i x j ) T (2) where F is the Frobenius inner product and M is an L L square matrix. To ensure that d M (x i, x j ) is a metric, the learned matrix M must be symmetric and PSD to guarantee that d M (x i, x j ) satisfies the symmetry, nonnegativity, and triangle inequality. However, the data may lie in a very high-dimensional space, which leads to a significant computational burden for solving M. To solve this problem of matrix M R L L, we can find a nonsquare matrix W R L D (D L) which defines a mapping function from the high-dimensional input space into a low-dimensional embedding and presents an alternative way to jointly perform dimensionality reduction and metric learning. According to this, we can rewrite d M (x i, x j ) [53] [56] to undertake the problem of dimensionality reduction as d M (x i, x j )= (x i x j ) T WW T (x i x j ) = (W T x i W T x j ) T (W T x i W T x j ) = W T x i W T x j. (3) The ultimate objective is to find an appropriate metric matrix W under the supervision of Λ and Ω, and the corresponding distance threshold b, such that the distance for (x i, x j ) Λ is smaller than the decision threshold b, and the distance for (x i, x j ) Ω is greater than the decision threshold b, so that the data points can be accurately classified. That is, the metric learning methods only lead to an absolute decision rule with a fixed threshold d. We can obtain the decision function as f fixed (d W )=d W (x i, x j ) b. (4) F A. Locally Adaptive Decision Constraints After the metric learning, we still need to make a decision for the sake of the classification. However, for data with complex distributions, the fixed threshold will be suboptimal even if the associated metric is correct. Thus, we adjust the decision constraints locally to relax the fixed threshold. By introducing the locally adaptive constraints, the local structures of the data can be used, which is the key to achieve a good classification performance. We use a locally adaptive decision function f(d ij ) to relax the fixed threshold, where d ii is the distance between a pair (x i, x j ) and is the reference to guide the changes. By using the locally adaptive constraints, we can guarantee that the greater the distance between similar pairs, the more f(d ii ) should shrink, while the smaller the distance between dissimilar pairs, the more f(d ij ) should expand. As a result, it allows us to make a decision by considering both the threshold b and the changes between the distances before and after metric learning. Based on this principle, we can form the locally adaptive decision constraints to compute the adaptive upper/lower bounds for (x i, x j ) as f Λ (d ij )=d ij (d (1/N Λ )/d ij c )(x i, x j ) Λ f Ω (d ij )=d ij +(d c /d (1/N Ω ) (6) )(x ij i, x j ) Ω where the constant d c 1. N Λ 1 and N Ω 1 are the scale factors that separately control the level of shrinkage and expansion. In this paper, we set d c = d max (where d max is the maximum distance between all the pairs). N Λ can be used to ensure that f Λ (d ij ) can shrink as fast as possible, which means that the smaller the value of N Λ, the faster f Λ (d ij ) will shrink. Meanwhile, N Ω guarantees the rapid expansion of f Ω (d ij ), which means that the larger the value of N Ω, the faster f Ω (d ij ) will expand. Clearly, we want to maximize the shrinkage and expansion of f(d ij ). Considering that N Λ < 1 cannot guarantee that the constraints are positive, we define N Λ =1to ensure that the locally adaptive constraints can shrink rapidly. Meanwhile, we define N Ω =1/ log(d c /(d c 2)) to ensure faster expansion, which allows us to distinguish the similar and dissimilar pairs by comparing them with d c. Finally, we obtain the locally adaptive decision constraints as follows: f Λ (d ij )= d ij (d ij /d max ) (x i, x j ) Λ f Ω (d ij )= d ij +(d max / N Ω dij ) (x i, x j ) Ω. (7) An illustration of the proposed locally adaptive decision constraints is shown in Fig. 1, which shows two pairs of samples from the hyperspectral image dataset. After the conventional metric learning, the Mahalanobis distance of a similar pair (the red points) is 80, which is higher than the threshold b =60, while the distance of a dissimilar pair (the green point and red point) is 40, which is lower than the threshold. With this decision paradigm, the two pairs may be misclassified. However, we can classify the pairs according to the locally adaptive decision constraints, with which the distance of the similar pair shrinks

6 6 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING Fig. 6. (a) Original scene of the Washington DC Mall image (channels 30, 90, and 150 for RGB) covering an area of pixels. (b) References of different classes. (c) Test area selected from the original data with a complex distribution. TABLE III NUMBERS OF SAMPLES OF THE WASHINGTON DC MALL DATASET USED FOR THEEXPERIMENT from 80 to 65, while the distance of the dissimilar pair expands from 40 to 55. Thus, it allows us to make a correct decision by considering both the threshold b and the changes between the distances before and after metric learning. B. Combining the Distance Metric and the Locally Adaptive Decision Constraints To acquire the metric matrix M (or W), the metric learning framework with a fixed threshold is as follows: min Ψ {M} + C L(ξ) M,ξ s.t. d M (x i, x j ) b(x i, x j ) Λ d M (x i, x j ) b(x i, x j ) Ω d M (x i, x j )=d M (x j, x i ),ξ 0, M 0 (8) where Ψ is the regularizer on M, and L is the loss term of ξ. ξ is used as the slack variable on the pairwise inequality constraint. With the help of the locally adaptive decision constraints, the proposed algorithm can adaptively determine the pairwise constraints, and can then effectively distinguish between the similar pairs and dissimilar pairs. Based on this rule, we can formulate the metric learning framework (8) as min Ψ(M)+C L(ξ) M,ξ s.t. d M (x i, x j ) f Λ (d ij )+ξ ij (x i, x j ) Λ d M (x i, x j ) f Ω (d ij ) ξ ij (x i, x j ) Ω d M (x i, x j )=d M (x j, x i ),ξ ij 0, M 0. (9) For (x i, x j ) Λ, wesety ij =1,f(d ij,y ij )=f Λ (d ij ), and y ij = 1,f(d ij,y ij )= f Ω (d ij ) for (x i, x j ) Ω. We can then simplify (9) as follows: min Ψ(M)+C L(ξ) M,ξ s.t. y ij d M (x i, x j ) f(d ij,y ij )+ξ ij ξ ij 0, M 0. (10) Some studies have shown the effectiveness of the squared F-norm regularizer and the good generalization performance of hinge loss [57], [58]. Thus, we adopt the squared F-norm regularizer and hinge loss in the proposed framework. Here, we generalize the above metric learning formulation to be a structured problem, which leads to a structured output instead of a simple label output. That is, a semidefinite programming problem needs to be solved because of the nonnegative constraint on the metric matrix M. To further solve the problem efficiently, we define a set of points U = {u ij } and u ij =(1;( (x i x j )(x i x j ) T )). By reforming (10) into an

7 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 7 TABLE IV INDIVIDUAL CLASS ACCURACIES,OAS, KAPPA STATISTIC VALUES, AND RUNNING TIMES (IN S) OF THE WASHINGTON DC MALL DATASET OBTAINED BY THEDIFFERENTCLASSIFICATIONMETHODS (dim = 60) Methods Class Original MNF-SVM SDA DLA LDA LFDA RCA NCA LMNN ITML MMML LADRml Road Grass Water Trails Trees Shadow Roof OA ±0.15 ±0.34 ±0.21 ±0.23 ±0.19 ±0.10 ±0.24 ±0.21 ±0.49 ±0.22 ±0.24 ±0.25 Kappa ± ± ± ± ± ± ± ± ± ± ± ± Time unbiased form, U forms the basis of the new input dataset points. Furthermore, we define ω =[f(d ij,y ij ); M], and then the formula can be rewritten as min M,ξ 1 2 ω 2 F + Cξ s.t. 1 n ωt c ijy ij u ij 1 ij n c ik ξ, ik c {0, 1} n,ξ 0, M 0. (11) We obtain a classifier which can be learned in an iterative manner in linear time with the above formulation, without exploiting more complex outputs. By using the metric learning framework, we can use both the global metric matrix M and the locally adaptive decision constraints to better distinguish between similar pairs and dissimilar pairs. However, the structured algorithm with PSD constraint demands exponential space and time, so it is a challenging problem to train the structured formulation on real large-scale data. Thus, we use the cutting plane algorithm to improve the objective function iteratively by using linear equalities. We start with an empty set of constraints, and we iteratively construct the locally adaptive decision constraints to optimize the solution. In each iteration, we calculate the unsatisfied locally adaptive decision constraints and combine them into the new constraint, and we then compute the optimum over the current metric matrix, in the case of the current compound constraints. A primal subgradient method is used to solve the optimization problem, which is reformulated into the following primal form: (ω) = argmin g(ω) = 1 ω 2 ω 2 F + Cξ n n s.t. 1 n c ijy ij (ω T u ij ) 1 ij n c ik ξ ik M 0, ω =[f(d ij,y ij ); M]. (12) We can then obtain the gradient of g(ω) with the correlative Lagrangian multipliers via ( g 1 ω = ω C ) n ωt c ijy ij u ij. (13) ij After solving the gradient step, we obtain a new ω and metric matrix M, which are defined as ω =[f(d ij,y ij ); M]. We then adopt spectral decomposition to eigen-decompose the metric matrix M by projecting the metric matrix M into the PSD cone to guarantee that the metric matrix is a PSD matrix. We start by setting the negative eigenvalues λ to zero, and we can then proceed with the eigen-decomposition with the corresponding eigenvector ν as m M = λ i νi T ν i. (14) i=1 We vectorize the new metric matrix M and place it in ω, and optimize until ω converges. We can then construct a linear projection matrix W R L D for the dimensionality reduction. For each test pixel vector x i R L D, we can calculate the reduced feature representation with the formula M = WW T x = W T x (15) which means that the original data x can be transformed in the low dimensionality of the Mahalanobis space. By using the new metric feature space, we can achieve classification. It is important to note that the proposed method is essentially different from our previous work (MMML) [42]. 1) First, the LADRml algorithm is proposed on the basis of data and classes with complex distributions in the hyperspectral image classification and dimensionality reduction problem, while the method proposed in [42] is aimed at hyperspectral target detection, considering that the target sample number is very low or the targets are difficult to detect compared with the huge background. 2) The LADRml algorithm adopts locally adaptive decision constraints to make a decision by considering both the decision threshold and the changes between the distances before and after metric learning. As a result, the local structures of the data can be used, which can solve the problem of data with complex distributions. However, the method proposed in [42] looks for a distance measure and makes a decision based on a fixed threshold, which is insufficient and suboptimal. When the between-class and within-class variations are complex, the method proposed in [42] cannot learn an effective metric for data with complex distributions. 3) The LADRml algorithm is a

8 8 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING Fig. 7. Classification maps for all the methods with the Washington DC Mall dataset. (a) Original, (b) MNF-SVM, (c) SDA, (d) DLA, (e) LDA, (f) LFDA, (g) RCA, (h) NCA, (i) LMNN, (j) ITML, (k) MMML, and (l) LADRml. local distance metric learning method that works by considering the locality of data distribution, which considers neighboring constraints and avoids adopting those conflicting constraints to further enhance the separability between different classes. In contrast, MMML is a global metric learning method, and it makes a decision based on a fixed threshold with pairwise constraints. 4) Both methods have strong generalization ability, but they differ with regard to parameters. The method proposed in [42] requires an extra parameter, i.e., the number of pairwise constraints for each class. Although the proposed algorithm has the same goal of learning a low-dimensional embedding of the data to improve the KNN classification, it should be noted that the LADRml algorithm is very different from LMNN in the following aspects. 1) LADRml and LMNN differ in the definition of the objective function. The proposed method adopts a structured formulation, which contains the locally adaptive decision constraints, instead of a simple label output. 2) From the constraints aspect, the proposed method learns a metric with pairwise constraints, while LMNN learns a metric with triplet constraints, which means that for each triplet (x i, x j, x k ) (where the class label of x i is the same as that of x j but different from that of x k, d M (x i, x j ) should be smaller than the distance d M (x i, x k ). Thus, the proposed method can work in more general cases by using the Euclidean distance of a pair to adaptively set the bounds. Moreover, LMNN adopts fixed bound-based constraints, and the proposed method adopts locally adaptive decision constraints, which makes LADRml more effective when the class variations are complex. 3) In terms of the optimization, the proposed method uses the cutting plane algorithm to

9 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 9 TABLE V NUMBERS OF SAMPLES OF THE ROSIS UNIVERSITY OF PAVIA DATASET USED FOR THEEXPERIMENT Fig. 8. Classification OA with regard to reduced dimensionality in the Washington DC Mall dataset. Fig. 9. (a) False-color composite image. (b) Ground-truth map. (c) Reference for the different classes from the ROSIS University of Pavia dataset. improve the objective function iteratively. 4) Finally, LMNN may have problems dealing with high-dimensional data, while the proposed method can achieve a stable performance. III. EXPERIMENTS AND DISCUSSION The proposed LADRml algorithm was evaluated on several popular hyperspectral imagery datasets, and we present the ex- perimental results demonstrating the benefits of the LADRml algorithm with the KNN classifier [59], [60], which is a classical classification method that is widely used in real-world applications. In order to provide a fair comparison, all the metric learning algorithms employed KNN as the classifier in the experiments. We compared the LADRml algorithm with four dimensionality reduction methods (SDA, DLA, LDA, and LFDA) and five representative metric learning methods (LMNN, NCA, RCA, ITML, and MMML), which can also be applied to dimensionality reduction. We used a 1-NN classifier [61] [63] for all the datasets. In addition, SVM with minimum noise fraction transform (MNF-SVM) [64] was also used as a comparison algorithm. The original dataset without metric learning (denoted as Original ) was also used and input into the 1-NN classifier. Similarly, for a fair comparison, all the methods used the same ground-truth data for all the experiments. For each dataset, we randomly selected 10% for each class as the training samples, and the rest were used as the test samples from the reference data to validate the performances. A. Hyperspectral Dataset Descriptions 1) The Indian Pines dataset, covering the Indian Pines region in Northwestern Indiana in 1992, was collected by the National Aeronautics and Spaces Administration s Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor. This scene comprises 220 spectral bands in the wavelength range from 0.4 to 2.5 μm with a size of The corresponding nominal spectral resolution is 10 nm, and the spatial resolution is approximately 20 m. In the experiment, we used a total of 200 radiance bands after removing some of the spectral bands affected by noise and water absorption. Due to the unbalanced number of available labeled pixels for each class, and the large number of mixed pixels in all the classes, it is a challenging image to classify. 2) The second airborne hyperspectral image is the Hyperspectral Digital Imagery Collection Experiment (HYDICE) Washington DC Mall dataset. This dataset contains 1280 lines, and each line has 307 pixels, including 210 bands within the 0.4- to 2.4-μm wavelength of the visible and infrared spectra. After discarding the water absorption channels, 191 channels remained. This dataset has significant variations between the spectra of the different classes, and is also a challenging image for classification.

10 10 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING TABLE VI INDIVIDUAL CLASS ACCURACIES,OAS, KAPPA STATISTIC VALUES, AND RUNNING TIMES (IN S) OF THE ROSIS UNIVERSITY OF PAVIA DATASET OBTAINED BY THE DIFFERENT CLASSIFICATION METHODS (dim = 70) Class Methods Original MNF-SVM SDA DLA LDA LFDA RCA NCA LMNN ITML MMML LADRml Asphalt Meadows Gravel Trees Metal sheets Bare soils Bitumen Bricks Shadows OA ± 0.25 ± 0.56 ± 0.18 ± 0.19 ± 0.74 ± 0.44 ± 0.24 ± 0.17 ± 0.16 ± 0.25 ± 0.23 ± 0.26 Kappa ± ± ± ± ± ± ± ± ± ± ± ± Time ) The third hyperspectral dataset was acquired by the airborne Reflective Optics System Imaging Spectrometer (ROSIS) sensor over the University of Pavia, Northern Italy. After removing 12 bands due to noise and water absorption, it comprises 103 spectral bands in the range from 0.41 to 0.82 μm. The image size in pixels is , with a high spatial resolution of 1.3 m per pixel, comprising complex urban, soil, vegetation, and other areas. The proposed algorithm has two parameters: the convergence value and the tradeoff parameter C. According to experience, the convergence value was set as The overall accuracy (OA) values of LADRml on the three datasets with respect to parameter C are presented in Fig. 2. From the figures, it can be observed that the results are relatively stable with regard to the values of C. This inspired us to set C = 1 in all the experiments. The classification accuracies were averaged over 10 runs for each method to reduce the possible bias induced by the random sampling. B. Hyperspectral Dataset Descriptions We first conducted an evaluation of the proposed algorithm with regard to the AVIRIS Indian Pines dataset. Fig. 3(a) shows the false-color image of the scene, while Fig. 3(b) shows the ground-truth classes of interest, which consist of 16 mutually exclusive classes with a total of samples. Fig. 3(c) shows the 1018 training samples, i.e., 10% of each class in the ground-truth image in all. The number of samples in each class is shown in Table I. This dataset is available online ( for testing the accuracies of hyperspectral image classification algorithms. Table II shows the individual class accuracies, the average OAs, and the average kappa statistic values with their standard deviations, obtained using 1018 training samples, as shown in Table I. From Table II, we can see that the LADRml algorithm performs significantly better than the RCA algorithm and the ITML algorithm in class accuracies, OA, and kappa statistic values. The Original, SDA, and LDA methods show a poorer discriminative ability than the LADRml algorithm. That is, the LADRml algorithm has strong generalization ability with limited training samples and can enhance the separability between different classes. The average running times for obtaining the distance metric in the scene are shown in the bottom line of the quantitative evaluation tables. For the running time comparison, it can be seen that DLA and LDA are the fastest methods, the proposed LADRml algorithm is comparable but slightly slower, and the other classical metric learning methods are the slowest. Overall, it is concluded that the proposed LADRml is more efficient than the other representation-based methods and obtains superior classification accuracy. In order to further show the good performance of the proposed LADRml algorithm, Fig. 4(a) (i) shows the classification maps along with the training samples, in which it can be observed that the proposed LADRml algorithm achieves the best classification result for most land-cover classes. To demonstrate the benefits of LADRml as a powerful dimensionality reduction tool for hyperspectral classification, we calculated the classification OA with regard to the reduced dimensionality (see Fig. 5). As shown in this figure, the LADRml algorithm obtains the best classification OA when dim (dimension) is larger than 15 and achieves a relatively stable result when dim is increased to a larger value, whereas DLA, MNF- SVM, and LFDA perform unsatisfactorily when dim reaches a large value. C. Experiment With the Washington DC Mall Dataset In this experiment, we adopted another challenging classification image which has significant spectral variation between different classes [65]. In order to show the classification results more efficiently, we used a subimage, selected from the original image shown in Fig. 6(a), as the experimental area (as shown in Fig. 6(c)), with a size of pixels. Fig. 6(b) gives the spectral variation of the different classes, and the colors of the lines represent the corresponding classes. In total, 5966 pixels from seven classes were used in the experiment. Table III details the samples of the Washington DC Mall dataset used in the experiment. As shown in Table III, we can see that the samples are

11 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 11 Fig. 10. Classification maps obtained from the ROSIS Pavia University dataset for: (a) Original, (b) MNF-SVM, (c) SDA, (d) DLA, (e) LDA, (f) LFDA, (g) RCA, (h) NCA, (i) LMNN, (j) ITML, (k) MMML, and (l) LADRml. Fig. 11. Two-dimensional representation of features for the different algorithms. (a) MNF-SVM, (b) SDA, (c) DLA, (d) LDA, (e) LFDA, (f) RCA, (g) NCA, (h) LMNN, (i) ITML, (j) MMML, and (k) LADRml.

12 12 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING Fig. 12. Classification OA with regard to reduced dimensionality in the ROSIS Pavia University dataset. limited, which makes the image difficult to classify. Similarly, all seven classes of test samples in the classification results were calculated by the ground truth with the same training and test sets for each round, producing the average results of ten independent rounds of classification reported in Table IV. As can be seen in Table IV, the proposed LADRml algorithm provides results which are comparable to those obtained by the other methods. Furthermore, the LADRml algorithm obtains the best OA and kappa in this dataset, and has more obvious advantages in most individual classes than the other methods. Fig. 7(a) (l) illustrates the classification maps of Original, MNF-SVM, SDA, DLA, LDA, LFDA, RCA, NCA, LMNN, ITML, MMML, and LADRml, respectively. Here, it can again be seen that the LADRml algorithm achieves the best classification performance for most classes. Furthermore, we enlarged the water class to better observe the classification performance, where it can be seen that the LADRml algorithm obtains fewer classification errors. This dataset has a complex distribution, but the proposed LADRml algorithm still achieves the best results. The locally adaptive decision constraints, added in the similarity measurement, allow for a more accurate measure of the relationship between the samples. Similarly, Fig. 8 indicates that the proposed method outperforms the other methods when dim is larger than 5 and achieves stability when the dimension increases to 10. D. Experiment With the ROSIS University of Pavia Dataset In this experiment, we adopted a classification image with a more complicated distribution, which also has larger areas. Fig. 9(a) shows the false-color composite of the image. Fig. 9(b) shows the nine ground-truth classes of interest with a total of samples, while 4273 samples were used for the training and samples for the testing. This image is also widely used in the hyperspectral image classification community [66], [67]. The samples used in the experiment are listed in Table V. As can be observed from Table V, the samples are huge and the problem of unbalance is more significant. Moreover, from the spectral variation of Fig. 9(c), we can see that most of the Fig. 13. Classification OAs of the different methods with different numbers of training samples in the three datasets. (a) AVIRIS Indian Pines dataset. (b) Washington DC Mall dataset. (c) ROSIS Pavia University dataset. spectra are quite similar, there are many different land-cover types, and the experimental image has a complex distribution. This results in the University of Pavia image being a difficult image for classification. Similarly, all nine classes of test samples were calculated by the ground truth in the classification results. The individual class

13 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 13 accuracies, average OAs, and average kappa statistic values with their standard deviations obtained in the algorithm comparison are listed in Table VI. The results reported in the table involved exactly the same number of training and test sets as those listed in Table V, allowing a fair comparison of the methods. As can be seen in Table VI, the proposed LADRml algorithm obtains the best class accuracy for most of the classes, and it also performs the best for the OA and kappa statistic values. Thus, we can conclude that the LADRml algorithm achieves the best classification result. As mentioned above, this image scene comprises data with a very complex distribution. In order to further show the good performance of the proposed algorithm, Fig. 10 shows the obtained classification maps along with the training samples. Here, it can be seen that the proposed LADRml algorithm outperforms the other algorithms, especially for bare soil and shadows, which have fewer training samples. The reason for this improvement may be that the LADRml algorithm can further enhance the separability between the different classes by applying the locally adaptive decision constraints. Fig. 11(a) (k) shows the sample distribution after application of the different methods, where it can be seen that RCA has the poorest discriminative ability. The reason for RCA s poor performance may be the assumption that it learns a globally linear transformation which may not be suitable for this data. LMNN and ITML have a weaker ability to separate the different classes. MNF-SVM, LDA, LFDA, NCA, SDA, and MMML can separate some classes well, while DLA results in some overlapping. The proposed LADRml algorithm can separate all of the classes well. In addition, the classification OAs (for all the methods) with an increase of the feature number dim are compared in Fig. 12. As can be seen in Fig. 12, the proposed algorithm achieves the best OA when dim is larger than 10, while the optimal OA stabilizes when the dimensionality increases to 15. The results allow a similar conclusion, i.e., LADRml outperforms the traditional dimensionality reduction methods and metric learning methods for dimensionality reduction in hyperspectral image classification. We also conducted experiments to investigate the influence of the number of training samples. We varied the amount of training data and studied the sensitivity of the proposed method relative to the conventional methods. Note that the training samples were the same in all cases. Fig. 13 shows the OA with the different numbers of training samples. For all the datasets, the training size was changed from 1% to 10% (note that 1% is the ratio of the number of training samples to the total labeled data). It is clear that the classification performance of LADRml is much better than the other methods. When the ratio of training samples becomes lower and lower, the performance of LADRml also decreases, but to a lesser degree than the other methods. This further confirms that the proposed LADRml is a competitive dimensionality reduction method, even with limited labeled data. IV. CONCLUSION In this paper, we have proposed the LADRml method, which combines global metric and locally adaptive decision constraints using a joint MMML model for hyperspectral image classification. The proposed LADRml algorithm can utilize the limited training samples and transfer the problem of dimensionality reduction without a certain distribution hypothesis into a MMML problem. Furthermore, the proposed LADRml algorithm adopts locally adaptive decision constraints to determine whether the pairs are similar or not by considering both the decision threshold b and the changes between the distances before and after metric learning. The experimental results show that the proposed LADRml algorithm performs better than the other state-of-the-art dimensionality reduction methods and metric learning methods on challenging hyperspectral datasets, i.e., the AVIRIS Indian Pines dataset, the HYDICE Washington DC Mall dataset, and the ROSIS University of Pavia dataset. For all these challenging datasets, the proposed LADRml method presents the highest accuracy. However, the processing time of the proposed algorithm is greater than the classical dimensionality reduction methods, although it is much less than the classical metric learning methods. In our future work, we will consider improved optimization methods and a flexible scheme to ensure the computational efficiency when learning optimal parameters. ACKNOWLEDGMENT The authors would like to sincerely thank Prof. D. Landgrebe for making the AVIRIS Indian Pines hyperspectral dataset available to the community, and Prof. P. Gamba for providing the ROSIS data from Pavia, Italy. They would also gratefully like to thank the handling editor and anonymous reviewers for their careful reading and helpful remarks, which significantly helped us to improve the technical quality and presentation of this paper. REFERENCES [1] F. Melgani and L. Bruzzone, Classification of hyperspectral remote sensing images with support vector machines, IEEE Trans. Geosci. Remote Sens., vol. 42, no. 8, pp , Aug [2] A. Plaza, P. Martinez, J. Plaza, and R. Perez, Dimensionality reduction and classification of hyperspectral image data using sequences of extended morphological transformations, IEEE Trans. Geosci. Remote Sens., vol. 43, no. 3, pp , Mar [3] R. Ji, Y. Gao, R. Hong, Q. Liu, D. Tao, and X. Li, Spectral-spatial constraint hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., vol. 52, no. 3, pp , Mar [4] J. Li, J. M. Bioucas-Dias, and A. Plaza, Spectral spatial classification of hyperspectral data using loopy belief propagation and active learning, IEEE Trans. Geosci. Remote Sens., vol. 51, no. 2, pp , Feb [5] J. C. Harsanyi and C.-I. Chang, Hyperspectral image classification and dimensionality reduction: An orthogonal subspace projection approach, IEEE Trans. Geosci. Remote Sens., vol. 32, pp , Jul [6] R. O. Green et al., Imaging spectroscopy and the airborne visible/infrared imaging spectrometer (AVIRIS), Remote Sens. Environ., vol. 65, no. 3, pp , Sep [7] C.-I. Chang, Hyperspectral Imaging: Techniques for Spectral Detection and Classification. New York, NY, USA: Kluwer, [8] D. Landgrebe, Hyperspectral image data analysis, IEEE Signal Process. Mag., vol. 19, no. 1, pp , Jan [9] G. Hughes, On the mean accuracy of statistical pattern recognizers, IEEE Trans. Inf. Theory, vol. IT-14, no. 1, pp , Jan [10] D. Landgrebe, Hyperspectral image data analysis as a high dimensional signal processing problems, IEEE Signal Process. Mag., vol. 19, no. 1, pp , Jan [11] W. Li, S. Prasad, J. E. Fowled, and L. M. Bruce, Locality-preserving dimensionality reduction and classification for hyperspectral image analysis, IEEE Trans. Geosci. Remote Sens., vol. 50, no. 4, pp , Apr

14 14 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING [12] Y. Zhou, J. Peng, and C. L. Chen, Dimension reduction using spatial and spectral regularized local discriminant embedding for hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., vol. 53, no. 2, pp , Feb [13] G. F. Hughes, On the mean accuracy of statistical pattern recognizers, IEEE Trans. Inf. Theory, vol. IT-14, no. 1, pp , Jan [14] L. M. Bruce, C. H. Koger, and J. Li, Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction, IEEE Trans. Geosci. Remote Sens., vol. 40, no. 10, pp , Oct [15] I. T. Jolliffe, Principal Component Analysis. New York, NY, USA: Springer, [16] S. T. Roweis and L. K. Saul, Nonlinear dimensionality reduction by locally linear embedding, Science, vol. 290, no. 22, pp ,2000. [17] C. Jutten and J. Hérault, Blind separation of sources, part I: An adaptive algorithm based on neuromimetic architecture, Signal Process., vol. 24, pp. 1 10, [18] X. He and P. Niyogi, Locality preserving projections, in Proc. Adv. Neural Inf. Process. Syst. Conf., 2003, pp [19] L. Clemmensen, T. Hastie, D. Witten, and B. Ersboll, Sparse discriminant analysis, Technometrics, vol. 53, no. 4, pp , Jan [20] T. Zhang, D. Tao, and J. Yang, Discriminative locality alignment, in Proc. Eur. Conf. Comput. Vis., 2008, pp [21] M. D. Farrell and R. M. Mersereau, On the impact of PCA dimension reduction for hyperspectral detection of difficult targets, IEEE Geosci. Remote Sens. Lett., vol. 2, no. 2, pp , Apr [22] J. Mairal, M. Elad, and G. Sapiro, Sparse representation for color image restoration, IEEE Trans. Image Process., vol. 17, no. 1, pp , Jan [23] J. Wright, Y. Ma, J. Mairal, G. Sapiro, T. Huang, and S. Yan, Sparse representation for computer vision and pattern recognition, Proc. IEEE, vol. 98, no. 6, pp , Jun [24] Y. Chen, N. M. Nasrabadi, and T. D. Tran, Hyperspectral image classification using dictionary-based sparse representation, IEEE Trans. Geosci. Remote Sens., vol. 49, no. 10, pp , Oct [25] O. Chapelle and A. Zien, Semi-supervised classification by low density separation, in Proc. 10th Int. Workshop Artif. Intell. Stat., 2005, pp [26] L. Bruzzone, M. Chi, and M. Marconcini, A novel transductive SVM for semisupervised classification of remote-sensing images, IEEE Trans. Geosci. Remote Sens., vol. 44, no. 11, pp , Nov [27] M. Chi and L. Bruzzone, Semisupervised classification of hyperspectral images by SVMs optimized in the primal, IEEE Trans. Geosci. Remote Sens., vol. 42, no. 6, pp , Jun [28] S. Rajan, J. Ghosh, and M. M. Crawford, An active learning approach to hyperspectral data classification, IEEE Trans. Geosci. Remote Sens., vol. 46, no. 4, pp , Apr [29] D. Tuia, F. Ratle, F. Pacifici, M. F. Kanevski, and W. J. Emery, Active learning methods for remote sensing image classification, IEEE Trans. Geosci. Remote Sens., vol. 47, no. 7, pp , Jul [30] S. J. Pan and Q. Yang, A survey on transfer learning, IEEE Trans. Knowl. Data Eng., vol. 20, no. 10, pp , Oct [31] C. Cortes and V. N. Vapnik, Support vector networks, Mach. Learn., vol. 20, no. 3, pp , Sep [32] J. A. Gualtieri and R. F. Cromp, Support vector machines for hyperspectral remote sensing classification, SPIE, vol. 3584, pp , Oct [33] F. Melgani and L. Bruzzone, Classification of hyperspectral remote sensing images with support vector machines, IEEE Trans. Geosci. Remote Sens., vol. 42, no. 8, pp , Aug [34] E. P. Xing, A. Y. Ng, M. I. Jordan, and S. Russell, Distance metric learning, with application to clustering with side-information, in Proc. Adv. Neural Inf. Process. Syst. Conf., 2002, pp [35] Y. Dong, B. Du, and L. Zhang, Target detection based on random forest metric learning, IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens.,vol.8, no. 4, pp , Apr [36] N. Shental, T. Hertz, D. Weinshall, and M. Pavel, Adjustment learning and relevant component analysis, in Proc. Eur. Conf. Comput. Vis.,2002, vol. 4, pp [37] J. Goldberger, S. Roweis, G. Hinton, and R. Salakhutdinov, Neighbourhood components analysis, in Proc. Adv. Neural Inf. Process. Syst. Conf., Dec. 2004, pp [38] K. Q. Weinberger, J. Blitzer, and L. K. Saul, Distance metric learning for large margin nearest neighbor classification, in Proc. 19th Annu. Conf. Neural Inf. Process. Syst., 2005, pp [39] J. V. Davis, B. Kulis, P. Jain, S. Sra, and I. S. Dhillon, Informationtheoretic metric learning, in Proc. 24th Int. Conf. Mach. Learn., Jun. 2007, pp [40] Q. Zhang, L. Zhang, Y. Yang, Y. Tian, and L. Weng, Local patch discriminative metric learning for hyperspectral image feature extraction, IEEE Geosci. Remote Sens. Lett., vol. 11, no. 3, pp , Mar [41] B. Bue, D. Thompson, M. Gilmore, and R. Castaño, Metric learning for hyperspectral image segmentation, in Proc. 3rd IEEE Workshop Hyperspectral Image Signal Process., Evol. Remote Sens., Jun. 2011, pp [42] Y. Dong, L. Zhang, L. Zhang, and B. Du, Maximum margin metric learning based target detection for hyperspectral images, ISPRS J. Photogramm. Remote Sens., vol. 108, pp , Oct [43] L. Zhang, L. Zhang, D. Tao, X. Huang, and B. Du, Hyperspectral remote sensing image subpixel target detection based on supervised metric learning, IEEE Trans. Geosci. Remote Sens., vol. 52, no. 8, pp , Aug [44] Z. Xu, K. Q. Weinberger, and O. Chapelle, Distance metric learning for kernel machines, arxiv preprint arxiv: , [45] C. Shen, J. Kim, and L. Wang, Scalable large-margin Mahalanobis distance metric learning, IEEE Trans. Neural Netw., vol. 21,no.9,pp , Sep [46] C. Xiong, D. M. Johnson, and J. J. Corso, Efficient max-margin metric learning, in Proc. 6th Int. Workshop Evol. Change Data Manage., 2012, pp [47] S. Shalev-Shwartz, Y. Singer, and N. Srebro, Pegasos: Primal estimated sub-gradient solver for SVM, in Proc. 24th Int. Conf. Mach. Learn.,2007, pp [48] V. Franc and S. Sonnenburg, Optimized cutting plane algorithm for support vector machines, in Proc. 25th Int. Conf. Mach. Learn., 2008, pp [49] T. Joachims, Training linear SVMs in linear time, in Proc. 12th ACM SIGKDD Int. Conf. Knowl. Disc. Data Min., 2006, pp [50] J. Lu, J. Hu, X. Zhou, and Y. Shang, Activity-based human identification using sparse coding and discriminative metric learning, in Proc. ACM Multimedia Conf., 2012, pp [51] J. Yu, D. Tao, J. Li, and J. Cheng, Semantic preserving distance metric learning and applications, Inf. Sci., vol. 281, pp , [52] F. Wang, Semisupervised metric learning by maximizing constraint margin, IEEE Trans. Syst., Man, Cybern. B, Cybern., vol. 41, no. 4, pp , Aug [53] F. Wang and C. Zhang, On discriminative semi-supervised classification, in Proc. Conf. Artif. Intell., 2008, pp [54] C. C. Chang, A boosting approach for supervised Mahalanobis distance metric learning, Pattern Recog., vol. 45, pp , [55] Z. Li, S. Chang, F. Liang, T. S. Huang, L. Cao, and J. R. Smith, Learning locally-adaptive decision functions for person verification, in Proc. IEEE Comput. Vis. Pattern Recog., 2013, pp [56] Q. Wang, W. Zuo, L. Zhang, and P. Li, Shrinkage expansion adaptive metric learning, in Proc. Eur. Conf. Comput. Vis., 2014, pp [57] B. Kulis, K. Saenko, and T. Darrell, What you saw is not what you get: Domain adaptation using asymmetric kernel transforms, in Proc. IEEE Comput. Vis. Pattern Recog., 2011, pp [58] Y. Luo, D. Tao, C. Xu, D. Li, and C. Xu, Vector-valued multi-view semisupervised learning for multi-label image classification, in Proc. 27th Conf. Artif. Intell., 2013, pp [59] T. M. Cover and P. E. Hart, Nearest neighbor pattern classification, IEEE Trans. Inf. Theory, vol. 13, no. 1, pp , Jan [60] H. He, S. Hawkins, W. Graco, and X. Yao, Application of genetic algorithm and k-nearest neighbour method in real world medical fraud detection problem, JACIII, vol. 4, no. 2, pp , [61] B. Du, L. Zhang, L. Zhang, T. Chen, and K. Wu, A discriminative manifold learning based dimension reduction method, Int. J. Fuzzy Syst., vol. 14, pp , Jun [62] Q. Shi, L. Zhang, and B. Du, Semisupervised discriminative locally enhanced alignment for hyperspectral image classification, IEEE Trans. Geosci. Remote Sens., vol. 51, no. 9, pp , Sep [63] B. D. Bue, An evaluation of low-rank Mahalanobis metric learning techniques for hyperspectral image classification, IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens., vol. 7, no. 4, pp , Apr [64] A. Green, M. Berman, P. Switzer, and M. Craig, A transformation for ordering multispectral data in terms of image quality with implications for noise removal, IEEE Trans. Geosci. Remote Sens., vol. 36, no. 1, pp , Jan

15 DONG et al.: EXPLORING LOCALLY ADAPTIVE DIMENSIONALITY REDUCTION FOR HYPERSPECTRAL IMAGE CLASSIFICATION 15 [65] L. Zhang, Q. Zhang, L. Zhang, D. Tao, X. Huang, and B. Du, Ensemble manifold regularized sparse low-rank approximation for multiview feature embedding, Pattern Recog., vol. 48, no. 10, pp , Oct [66] Y. Gu and K. Feng, Optimized Laplacian SVM with distance metric learning for hyperspectral image classification, IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens., vol. 6, no. 3, pp , Jun [67] W. Li and Q. Du, Joint within-class collaborative representation for hyperspectral image classification, IEEE J. Sel. Topics Appl. Earth Obs. Remote Sens., vol. 52, no. 6, pp , Jun Yanni Dong (S 14) received the B.S. degree in sciences and techniques of remote sensing from Wuhan University, Wuhan, China, in 2012, where she is currently working toward the Ph.D. degree in the State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing. Her current research interests include pattern recognition in remote sensing images, hyperspectral image processing, and metric learning. Bo Du (M 10 SM 15) received the B.S. degree from Wuhan University, Wuhan, China, in 2005, where he receive the Ph.D. degree in photogrammetry and remote sensing from the State Key Lab of Information Engineering in Surveying, Mapping, and Remote Sensing in He is currently a Professor with the School of Computer, Wuhan University. His major research interests include pattern recognition, hyperspectral image processing, and signal processing. Liangpei Zhang (M 06 SM 08) received the B.S. degree in physics from Hunan Normal University, Changsha, China, in 1982, the M.S. degree in optics from Chinese Academy of Sciences, Xian, China, in 1988, and the Ph.D. degree in photogrammetry and remote sensing from Wuhan University, Wuhan, China, in He is currently the Head of the Remote Sensing Division, State Key Laboratory of Information Engineering in Surveying, Mapping, and Remote Sensing, Wuhan University. He is also a Chang-Jiang Scholar Chair Professor appointed by the Ministry of Education of China. He is currently a Principal Scientist for the China State Key Basic Research Project ( ) appointed by the Ministry of National Science and Technology of China to lead the remote sensing program in China. He is the holder of 15 patents. He has more than 360 research papers. His research interests include hyperspectral remote sensing, high-resolution remote sensing, image processing, and artificial intelligence. Dr. Zhang is a Fellow of the Institution of Engineering and Technology, an Executive Member (Board of Governor) of the China National Committee of the International Geosphere-Biosphere Programme, and an Executive Member of the China Society of Image and Graphics. He regularly serves as a Cochair of the series SPIE Conferences on Multispectral Image Processing and Pattern Recognition, Conference on Asia Remote Sensing, and many other conferences. He edits several conference proceedings, issues, and Geoinformatics symposiums. He also serves as an Associate Editor of the International Journal of Ambient Computing and Intelligence, the International Journal of Image and Graphics, the International Journal of Digital Multimedia Broadcasting, thejournal of Geo-spatial Information Science,theJournal of Remote Sensing, and the IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING. Lefei Zhang (S 11 M 14) received the B.S. and Ph.D. degrees from Wuhan University, Wuhan, China, in 2008 and 2013, respectively. From August 2013 to July 2015, he was with the School of Computer, Wuhan University, as a Postdoctoral Researcher, and he was a Visiting Scholar with the CAD & CG Lab, Zhejiang University in 2015, and a Big Data Institute Visitor in the Department of Statistical Science, University College London in He is currently a Lecturer with the School of Computer, Wuhan University, and also a Hong Kong Scholar in the Department of Computing, The Hong Kong Polytechnic University, Kowloon, Hong Kong. His research interests include pattern recognition, image processing, and remote sensing. Dr. Zhang is a Reviewer of more than 20 international journals, including the IEEE TIP, the TNNLS, the TMM, and the TGRS.

Classification of Hyperspectral Data over Urban. Areas Using Directional Morphological Profiles and. Semi-supervised Feature Extraction

Classification of Hyperspectral Data over Urban. Areas Using Directional Morphological Profiles and. Semi-supervised Feature Extraction IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL.X, NO.X, Y 1 Classification of Hyperspectral Data over Urban Areas Using Directional Morphological Profiles and Semi-supervised

More information

Hyperspectral and Multispectral Image Fusion Using Local Spatial-Spectral Dictionary Pair

Hyperspectral and Multispectral Image Fusion Using Local Spatial-Spectral Dictionary Pair Hyperspectral and Multispectral Image Fusion Using Local Spatial-Spectral Dictionary Pair Yifan Zhang, Tuo Zhao, and Mingyi He School of Electronics and Information International Center for Information

More information

Hyperspectral Image Classification Using Gradient Local Auto-Correlations

Hyperspectral Image Classification Using Gradient Local Auto-Correlations Hyperspectral Image Classification Using Gradient Local Auto-Correlations Chen Chen 1, Junjun Jiang 2, Baochang Zhang 3, Wankou Yang 4, Jianzhong Guo 5 1. epartment of Electrical Engineering, University

More information

Spectral-Spatial Response for Hyperspectral Image Classification

Spectral-Spatial Response for Hyperspectral Image Classification Article Spectral-Spatial Response for Hyperspectral Image Classification Yantao Wei 1,2, *,, Yicong Zhou 2, and Hong Li 3 1 School of Educational Information Technology, Central China Normal University,

More information

Does Normalization Methods Play a Role for Hyperspectral Image Classification?

Does Normalization Methods Play a Role for Hyperspectral Image Classification? Does Normalization Methods Play a Role for Hyperspectral Image Classification? Faxian Cao 1, Zhijing Yang 1*, Jinchang Ren 2, Mengying Jiang 1, Wing-Kuen Ling 1 1 School of Information Engineering, Guangdong

More information

A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2

A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2 A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA Naoto Yokoya 1 and Akira Iwasaki 1 Graduate Student, Department of Aeronautics and Astronautics, The University of

More information

Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image

Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image Murinto 1, Nur Rochmah Dyah PA 2 1,2 Department of Informatics Engineering

More information

HYPERSPECTRAL image (HSI) acquired by spaceborne

HYPERSPECTRAL image (HSI) acquired by spaceborne 1 SuperPCA: A Superpixelwise PCA Approach for Unsupervised Feature Extraction of Hyperspectral Imagery Junjun Jiang, Member, IEEE, Jiayi Ma, Member, IEEE, Chen Chen, Member, IEEE, Zhongyuan Wang, Member,

More information

DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION

DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION S.Dhanalakshmi #1 #PG Scholar, Department of Computer Science, Dr.Sivanthi Aditanar college of Engineering, Tiruchendur

More information

Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data

Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data Wenzhi liao a, *, Frieke Van Coillie b, Flore Devriendt b, Sidharta Gautama a, Aleksandra Pizurica

More information

GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION

GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION Nasehe Jamshidpour a, Saeid Homayouni b, Abdolreza Safari a a Dept. of Geomatics Engineering, College of Engineering,

More information

Hyperspectral Image Classification by Using Pixel Spatial Correlation

Hyperspectral Image Classification by Using Pixel Spatial Correlation Hyperspectral Image Classification by Using Pixel Spatial Correlation Yue Gao and Tat-Seng Chua School of Computing, National University of Singapore, Singapore {gaoyue,chuats}@comp.nus.edu.sg Abstract.

More information

HYPERSPECTRAL imagery (HSI) records hundreds of

HYPERSPECTRAL imagery (HSI) records hundreds of IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 11, NO. 1, JANUARY 2014 173 Classification Based on 3-D DWT and Decision Fusion for Hyperspectral Image Analysis Zhen Ye, Student Member, IEEE, Saurabh

More information

Remote Sensed Image Classification based on Spatial and Spectral Features using SVM

Remote Sensed Image Classification based on Spatial and Spectral Features using SVM RESEARCH ARTICLE OPEN ACCESS Remote Sensed Image Classification based on Spatial and Spectral Features using SVM Mary Jasmine. E PG Scholar Department of Computer Science and Engineering, University College

More information

IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 12, NO. 2, FEBRUARY

IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 12, NO. 2, FEBRUARY IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 12, NO. 2, FEBRUARY 2015 349 Subspace-Based Support Vector Machines for Hyperspectral Image Classification Lianru Gao, Jun Li, Member, IEEE, Mahdi Khodadadzadeh,

More information

STRATIFIED SAMPLING METHOD BASED TRAINING PIXELS SELECTION FOR HYPER SPECTRAL REMOTE SENSING IMAGE CLASSIFICATION

STRATIFIED SAMPLING METHOD BASED TRAINING PIXELS SELECTION FOR HYPER SPECTRAL REMOTE SENSING IMAGE CLASSIFICATION Volume 117 No. 17 2017, 121-126 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu ijpam.eu STRATIFIED SAMPLING METHOD BASED TRAINING PIXELS SELECTION FOR HYPER

More information

High-Resolution Image Classification Integrating Spectral-Spatial-Location Cues by Conditional Random Fields

High-Resolution Image Classification Integrating Spectral-Spatial-Location Cues by Conditional Random Fields IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 25, NO. 9, SEPTEMBER 2016 4033 High-Resolution Image Classification Integrating Spectral-Spatial-Location Cues by Conditional Random Fields Ji Zhao, Student

More information

HYPERSPECTRAL imagery has been increasingly used

HYPERSPECTRAL imagery has been increasingly used IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 5, MAY 2017 597 Transferred Deep Learning for Anomaly Detection in Hyperspectral Imagery Wei Li, Senior Member, IEEE, Guodong Wu, and Qian Du, Senior

More information

IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 13, NO. 8, AUGUST

IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 13, NO. 8, AUGUST IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 13, NO. 8, AUGUST 2016 1059 A Modified Locality-Preserving Projection Approach for Hyperspectral Image Classification Yongguang Zhai, Lifu Zhang, Senior

More information

Hyperspectral Image Classification via Kernel Sparse Representation

Hyperspectral Image Classification via Kernel Sparse Representation 1 Hyperspectral Image Classification via Kernel Sparse Representation Yi Chen 1, Nasser M. Nasrabadi 2, Fellow, IEEE, and Trac D. Tran 1, Senior Member, IEEE 1 Department of Electrical and Computer Engineering,

More information

Spectral-spatial rotation forest for hyperspectral image classification

Spectral-spatial rotation forest for hyperspectral image classification Spectral-spatial rotation forest for hyperspectral image classification Junshi Xia, Lionel Bombrun, Yannick Berthoumieu, Christian Germain, Peijun Du To cite this version: Junshi Xia, Lionel Bombrun, Yannick

More information

Lab 9. Julia Janicki. Introduction

Lab 9. Julia Janicki. Introduction Lab 9 Julia Janicki Introduction My goal for this project is to map a general land cover in the area of Alexandria in Egypt using supervised classification, specifically the Maximum Likelihood and Support

More information

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 24, NO. 7, JULY

IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 24, NO. 7, JULY IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 24, NO. 7, JULY 2015 2037 Spatial Coherence-Based Batch-Mode Active Learning for Remote Sensing Image Classification Qian Shi, Bo Du, Member, IEEE, and Liangpei

More information

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference

Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Detecting Burnscar from Hyperspectral Imagery via Sparse Representation with Low-Rank Interference Minh Dao 1, Xiang Xiang 1, Bulent Ayhan 2, Chiman Kwan 2, Trac D. Tran 1 Johns Hopkins Univeristy, 3400

More information

1314 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 4, APRIL 2014

1314 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 4, APRIL 2014 1314 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 4, APRIL 2014 A Support Vector Conditional Random Fields Classifier With a Mahalanobis Distance Boundary

More information

THE detailed spectral information of hyperspectral

THE detailed spectral information of hyperspectral 1358 IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 14, NO. 8, AUGUST 2017 Locality Sensitive Discriminant Analysis for Group Sparse Representation-Based Hyperspectral Imagery Classification Haoyang

More information

Spectral Angle Based Unary Energy Functions for Spatial-Spectral Hyperspectral Classification Using Markov Random Fields

Spectral Angle Based Unary Energy Functions for Spatial-Spectral Hyperspectral Classification Using Markov Random Fields Rochester Institute of Technology RIT Scholar Works Presentations and other scholarship 7-31-2016 Spectral Angle Based Unary Energy Functions for Spatial-Spectral Hyperspectral Classification Using Markov

More information

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 10, OCTOBER

IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 10, OCTOBER IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 10, OCTOBER 2015 5677 Domain Adaptation for Remote Sensing Image Classification: A Low-Rank Reconstruction and Instance Weighting Label

More information

Exploring Structural Consistency in Graph Regularized Joint Spectral-Spatial Sparse Coding for Hyperspectral Image Classification

Exploring Structural Consistency in Graph Regularized Joint Spectral-Spatial Sparse Coding for Hyperspectral Image Classification 1 Exploring Structural Consistency in Graph Regularized Joint Spectral-Spatial Sparse Coding for Hyperspectral Image Classification Changhong Liu, Jun Zhou, Senior Member, IEEE, Jie Liang, Yuntao Qian,

More information

PoS(CENet2017)005. The Classification of Hyperspectral Images Based on Band-Grouping and Convolutional Neural Network. Speaker.

PoS(CENet2017)005. The Classification of Hyperspectral Images Based on Band-Grouping and Convolutional Neural Network. Speaker. The Classification of Hyperspectral Images Based on Band-Grouping and Convolutional Neural Network 1 Xi an Hi-Tech Institute Xi an 710025, China E-mail: dr-f@21cnl.c Hongyang Gu Xi an Hi-Tech Institute

More information

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method

Face Recognition Based on LDA and Improved Pairwise-Constrained Multiple Metric Learning Method Journal of Information Hiding and Multimedia Signal Processing c 2016 ISSN 2073-4212 Ubiquitous International Volume 7, Number 5, September 2016 Face Recognition ased on LDA and Improved Pairwise-Constrained

More information

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality

More information

c 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all

c 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all c 2017 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising

More information

Learning Compact and Effective Distance Metrics with Diversity Regularization. Pengtao Xie. Carnegie Mellon University

Learning Compact and Effective Distance Metrics with Diversity Regularization. Pengtao Xie. Carnegie Mellon University Learning Compact and Effective Distance Metrics with Diversity Regularization Pengtao Xie Carnegie Mellon University 1 Distance Metric Learning Similar Dissimilar Distance Metric Wide applications in retrieval,

More information

Applying Supervised Learning

Applying Supervised Learning Applying Supervised Learning When to Consider Supervised Learning A supervised learning algorithm takes a known set of input data (the training set) and known responses to the data (output), and trains

More information

Hyperspectral Remote Sensing

Hyperspectral Remote Sensing Hyperspectral Remote Sensing Multi-spectral: Several comparatively wide spectral bands Hyperspectral: Many (could be hundreds) very narrow spectral bands GEOG 4110/5100 30 AVIRIS: Airborne Visible/Infrared

More information

Textural Features for Hyperspectral Pixel Classification

Textural Features for Hyperspectral Pixel Classification Textural Features for Hyperspectral Pixel Classification Olga Rajadell, Pedro García-Sevilla, and Filiberto Pla Depto. Lenguajes y Sistemas Informáticos Jaume I University, Campus Riu Sec s/n 12071 Castellón,

More information

REMOTE sensing hyperspectral images (HSI) are acquired

REMOTE sensing hyperspectral images (HSI) are acquired IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 10, NO. 3, MARCH 2017 1151 Exploring Structural Consistency in Graph Regularized Joint Spectral-Spatial Sparse Coding

More information

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1 IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1 GSEAD: Graphical Scoring Estimation for Hyperspectral Anomaly Detection Rui Zhao, Student Member, IEEE, and Liangpei Zhang,

More information

Title: A Deep Network Architecture for Super-resolution aided Hyperspectral Image Classification with Class-wise Loss

Title: A Deep Network Architecture for Super-resolution aided Hyperspectral Image Classification with Class-wise Loss 2018 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising

More information

DIMENSION REDUCTION FOR HYPERSPECTRAL DATA USING RANDOMIZED PCA AND LAPLACIAN EIGENMAPS

DIMENSION REDUCTION FOR HYPERSPECTRAL DATA USING RANDOMIZED PCA AND LAPLACIAN EIGENMAPS DIMENSION REDUCTION FOR HYPERSPECTRAL DATA USING RANDOMIZED PCA AND LAPLACIAN EIGENMAPS YIRAN LI APPLIED MATHEMATICS, STATISTICS AND SCIENTIFIC COMPUTING ADVISOR: DR. WOJTEK CZAJA, DR. JOHN BENEDETTO DEPARTMENT

More information

Learning with infinitely many features

Learning with infinitely many features Learning with infinitely many features R. Flamary, Joint work with A. Rakotomamonjy F. Yger, M. Volpi, M. Dalla Mura, D. Tuia Laboratoire Lagrange, Université de Nice Sophia Antipolis December 2012 Example

More information

KERNEL-based methods, such as support vector machines

KERNEL-based methods, such as support vector machines 48 IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 12, NO. 1, JANUARY 2015 Kernel Collaborative Representation With Tikhonov Regularization for Hyperspectral Image Classification Wei Li, Member, IEEE,QianDu,Senior

More information

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER A.Shabbir 1, 2 and G.Verdoolaege 1, 3 1 Department of Applied Physics, Ghent University, B-9000 Ghent, Belgium 2 Max Planck Institute

More information

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2 A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation Kwanyong Lee 1 and Hyeyoung Park 2 1. Department of Computer Science, Korea National Open

More information

HYPERSPECTRAL remote sensing sensors provide hundreds

HYPERSPECTRAL remote sensing sensors provide hundreds 70 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 1, JANUARY 2015 Spectral Spatial Classification of Hyperspectral Data via Morphological Component Analysis-Based Image Separation Zhaohui

More information

A Novel Clustering-Based Feature Representation for the Classification of Hyperspectral Imagery

A Novel Clustering-Based Feature Representation for the Classification of Hyperspectral Imagery Remote Sens. 2014, 6, 5732-5753; doi:10.3390/rs6065732 Article OPEN ACCESS remote sensing ISSN 2072-4292 www.mdpi.com/journal/remotesensing A Novel Clustering-Based Feature Representation for the Classification

More information

4202 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 8, AUGUST 2015

4202 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 8, AUGUST 2015 4202 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 8, AUGUST 2015 Adaptive Multiobjective Memetic Fuzzy Clustering Algorithm for Remote Sensing Imagery Ailong Ma, Graduate Student Member,

More information

THE imaging spectrometer, airborne or spaceborne, is a

THE imaging spectrometer, airborne or spaceborne, is a IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING 1 Semisupervised Discriminative Locally Enhanced Alignment for Hyperspectral Image Classification Qian Shi, Student Member, IEEE, Liangpei Zhang, Senior

More information

A Robust Band Compression Technique for Hyperspectral Image Classification

A Robust Band Compression Technique for Hyperspectral Image Classification A Robust Band Compression Technique for Hyperspectral Image Classification Qazi Sami ul Haq,Lixin Shi,Linmi Tao,Shiqiang Yang Key Laboratory of Pervasive Computing, Ministry of Education Department of

More information

HYPERSPECTRAL sensors provide a rich source of

HYPERSPECTRAL sensors provide a rich source of Fast Hyperspectral Feature Reduction Using Piecewise Constant Function Approximations Are C. Jensen, Student member, IEEE and Anne Schistad Solberg, Member, IEEE Abstract The high number of spectral bands

More information

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE 2014 2147 Automatic Framework for Spectral Spatial Classification Based on Supervised Feature Extraction

More information

Fuzzy Entropy based feature selection for classification of hyperspectral data

Fuzzy Entropy based feature selection for classification of hyperspectral data Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering NIT Kurukshetra, 136119 mpce_pal@yahoo.co.uk Abstract: This paper proposes to use

More information

R-VCANet: A New Deep Learning-Based Hyperspectral Image Classification Method

R-VCANet: A New Deep Learning-Based Hyperspectral Image Classification Method IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1 R-VCANet: A New Deep Learning-Based Hyperspectral Image Classification Method Bin Pan, Zhenwei Shi and Xia Xu Abstract

More information

sensors ISSN

sensors ISSN Sensors 2009, 9, 196-218; doi:10.3390/s90100196 OPEN ACCESS sensors ISSN 1424-8220 www.mdpi.com/journal/sensors Article Multi-Channel Morphological Profiles for Classification of Hyperspectral Images Using

More information

HYPERSPECTRAL REMOTE SENSING

HYPERSPECTRAL REMOTE SENSING HYPERSPECTRAL REMOTE SENSING By Samuel Rosario Overview The Electromagnetic Spectrum Radiation Types MSI vs HIS Sensors Applications Image Analysis Software Feature Extraction Information Extraction 1

More information

MULTI/HYPERSPECTRAL imagery has the potential to

MULTI/HYPERSPECTRAL imagery has the potential to IEEE GEOSCIENCE AND REMOTE SENSING ETTERS, VO. 11, NO. 12, DECEMBER 2014 2183 Three-Dimensional Wavelet Texture Feature Extraction and Classification for Multi/Hyperspectral Imagery Xian Guo, Xin Huang,

More information

Schroedinger Eigenmaps with Nondiagonal Potentials for Spatial-Spectral Clustering of Hyperspectral Imagery

Schroedinger Eigenmaps with Nondiagonal Potentials for Spatial-Spectral Clustering of Hyperspectral Imagery Schroedinger Eigenmaps with Nondiagonal Potentials for Spatial-Spectral Clustering of Hyperspectral Imagery Nathan D. Cahill a, Wojciech Czaja b, and David W. Messinger c a Center for Applied and Computational

More information

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo

CSE 6242 A / CS 4803 DVA. Feb 12, Dimension Reduction. Guest Lecturer: Jaegul Choo CSE 6242 A / CS 4803 DVA Feb 12, 2013 Dimension Reduction Guest Lecturer: Jaegul Choo CSE 6242 A / CS 4803 DVA Feb 12, 2013 Dimension Reduction Guest Lecturer: Jaegul Choo Data is Too Big To Do Something..

More information

HYPERSPECTRAL remote sensing images are very important

HYPERSPECTRAL remote sensing images are very important 762 IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 6, NO. 4, OCTOBER 2009 Ensemble Classification Algorithm for Hyperspectral Remote Sensing Data Mingmin Chi, Member, IEEE, Qian Kun, Jón Atli Beneditsson,

More information

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1

Cluster Analysis. Mu-Chun Su. Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Cluster Analysis Mu-Chun Su Department of Computer Science and Information Engineering National Central University 2003/3/11 1 Introduction Cluster analysis is the formal study of algorithms and methods

More information

A Multichannel Gray Level Co-Occurrence Matrix for Multi/Hyperspectral Image Texture Representation

A Multichannel Gray Level Co-Occurrence Matrix for Multi/Hyperspectral Image Texture Representation Remote Sens. 2014, 6, 8424-8445; doi:10.3390/rs6098424 Article OPEN ACCESS remote sensing ISSN 2072-4292 www.mdpi.com/journal/remotesensing A Multichannel Gray Level Co-Occurrence Matrix for Multi/Hyperspectral

More information

The Comparative Study of Machine Learning Algorithms in Text Data Classification*

The Comparative Study of Machine Learning Algorithms in Text Data Classification* The Comparative Study of Machine Learning Algorithms in Text Data Classification* Wang Xin School of Science, Beijing Information Science and Technology University Beijing, China Abstract Classification

More information

Dimension Reduction CS534

Dimension Reduction CS534 Dimension Reduction CS534 Why dimension reduction? High dimensionality large number of features E.g., documents represented by thousands of words, millions of bigrams Images represented by thousands of

More information

FACE RECOGNITION USING SUPPORT VECTOR MACHINES

FACE RECOGNITION USING SUPPORT VECTOR MACHINES FACE RECOGNITION USING SUPPORT VECTOR MACHINES Ashwin Swaminathan ashwins@umd.edu ENEE633: Statistical and Neural Pattern Recognition Instructor : Prof. Rama Chellappa Project 2, Part (b) 1. INTRODUCTION

More information

Spatially variant dimensionality reduction for the visualization of multi/hyperspectral images

Spatially variant dimensionality reduction for the visualization of multi/hyperspectral images Author manuscript, published in "International Conference on Image Analysis and Recognition, Burnaby : Canada (2011)" DOI : 10.1007/978-3-642-21593-3_38 Spatially variant dimensionality reduction for the

More information

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM

CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM 96 CHAPTER 6 IDENTIFICATION OF CLUSTERS USING VISUAL VALIDATION VAT ALGORITHM Clustering is the process of combining a set of relevant information in the same group. In this process KM algorithm plays

More information

Spectral Classification

Spectral Classification Spectral Classification Spectral Classification Supervised versus Unsupervised Classification n Unsupervised Classes are determined by the computer. Also referred to as clustering n Supervised Classes

More information

Face Recognition via Sparse Representation

Face Recognition via Sparse Representation Face Recognition via Sparse Representation John Wright, Allen Y. Yang, Arvind, S. Shankar Sastry and Yi Ma IEEE Trans. PAMI, March 2008 Research About Face Face Detection Face Alignment Face Recognition

More information

Semi-Supervised Clustering with Partial Background Information

Semi-Supervised Clustering with Partial Background Information Semi-Supervised Clustering with Partial Background Information Jing Gao Pang-Ning Tan Haibin Cheng Abstract Incorporating background knowledge into unsupervised clustering algorithms has been the subject

More information

Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification

Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification Selim Aksoy and H. Gökhan Akçay Bilkent University Department of Computer Engineering Bilkent, 06800, Ankara, Turkey

More information

HYPERSPECTRAL images (HSIs), spanning the visible

HYPERSPECTRAL images (HSIs), spanning the visible 5338 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 53, NO. 10, OCTOBER 2015 Efficient Superpixel-Level Multitask Joint Sparse Representation for Hyperspectral Image Classification Jiayi Li,

More information

10-701/15-781, Fall 2006, Final

10-701/15-781, Fall 2006, Final -7/-78, Fall 6, Final Dec, :pm-8:pm There are 9 questions in this exam ( pages including this cover sheet). If you need more room to work out your answer to a question, use the back of the page and clearly

More information

Convex combination of adaptive filters for a variable tap-length LMS algorithm

Convex combination of adaptive filters for a variable tap-length LMS algorithm Loughborough University Institutional Repository Convex combination of adaptive filters for a variable tap-length LMS algorithm This item was submitted to Loughborough University's Institutional Repository

More information

Discriminant Analysis-Based Dimension Reduction for Hyperspectral Image Classification

Discriminant Analysis-Based Dimension Reduction for Hyperspectral Image Classification Satellite View istockphoto.com/frankramspott puzzle outline footage firm, inc. Discriminant Analysis-Based Dimension Reduction for Hyperspectral Image Classification A survey of the most recent advances

More information

Comparison of Support Vector Machine-Based Processing Chains for Hyperspectral Image Classification

Comparison of Support Vector Machine-Based Processing Chains for Hyperspectral Image Classification Comparison of Support Vector Machine-Based Processing Chains for Hyperspectral Image Classification Marta Rojas a, Inmaculada Dópido a, Antonio Plaza a and Paolo Gamba b a Hyperspectral Computing Laboratory

More information

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS

CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS 38 CHAPTER 3 PRINCIPAL COMPONENT ANALYSIS AND FISHER LINEAR DISCRIMINANT ANALYSIS 3.1 PRINCIPAL COMPONENT ANALYSIS (PCA) 3.1.1 Introduction In the previous chapter, a brief literature review on conventional

More information

Spectral-Spatial Classification of Hyperspectral Images Using Approximate Sparse Multinomial Logistic Regression

Spectral-Spatial Classification of Hyperspectral Images Using Approximate Sparse Multinomial Logistic Regression Spectral-Spatial Classification of Hyperspectral Images Using Approximate Sparse Multinomial Logistic Regression KORAY KAYABOL Dept. of Electronics Engineering Gebze Technical University Kocaeli TURKEY

More information

Unsupervised Learning

Unsupervised Learning Networks for Pattern Recognition, 2014 Networks for Single Linkage K-Means Soft DBSCAN PCA Networks for Kohonen Maps Linear Vector Quantization Networks for Problems/Approaches in Machine Learning Supervised

More information

Technical Report. Title: Manifold learning and Random Projections for multi-view object recognition

Technical Report. Title: Manifold learning and Random Projections for multi-view object recognition Technical Report Title: Manifold learning and Random Projections for multi-view object recognition Authors: Grigorios Tsagkatakis 1 and Andreas Savakis 2 1 Center for Imaging Science, Rochester Institute

More information

Feature Selection for Classification of Remote Sensed Hyperspectral Images: A Filter approach using Genetic Algorithm and Cluster Validity

Feature Selection for Classification of Remote Sensed Hyperspectral Images: A Filter approach using Genetic Algorithm and Cluster Validity Feature Selection for Classification of Remote Sensed Hyperspectral Images: A Filter approach using Genetic Algorithm and Cluster Validity A. B. Santos 1, C. S. F. de S. Celes 1, A. de A. Araújo 1, and

More information

Application of Support Vector Machine Algorithm in Spam Filtering

Application of Support Vector Machine Algorithm in  Spam Filtering Application of Support Vector Machine Algorithm in E-Mail Spam Filtering Julia Bluszcz, Daria Fitisova, Alexander Hamann, Alexey Trifonov, Advisor: Patrick Jähnichen Abstract The problem of spam classification

More information

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION

CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION CHAPTER 6 MODIFIED FUZZY TECHNIQUES BASED IMAGE SEGMENTATION 6.1 INTRODUCTION Fuzzy logic based computational techniques are becoming increasingly important in the medical image analysis arena. The significant

More information

Metric Learning Applied for Automatic Large Image Classification

Metric Learning Applied for Automatic Large Image Classification September, 2014 UPC Metric Learning Applied for Automatic Large Image Classification Supervisors SAHILU WENDESON / IT4BI TOON CALDERS (PhD)/ULB SALIM JOUILI (PhD)/EuraNova Image Database Classification

More information

Introduction to digital image classification

Introduction to digital image classification Introduction to digital image classification Dr. Norman Kerle, Wan Bakx MSc a.o. INTERNATIONAL INSTITUTE FOR GEO-INFORMATION SCIENCE AND EARTH OBSERVATION Purpose of lecture Main lecture topics Review

More information

PRINCIPAL components analysis (PCA) is a widely

PRINCIPAL components analysis (PCA) is a widely 1586 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 44, NO. 6, JUNE 2006 Independent Component Analysis-Based Dimensionality Reduction With Applications in Hyperspectral Image Analysis Jing Wang,

More information

PARALLEL IMPLEMENTATION OF MORPHOLOGICAL PROFILE BASED SPECTRAL-SPATIAL CLASSIFICATION SCHEME FOR HYPERSPECTRAL IMAGERY

PARALLEL IMPLEMENTATION OF MORPHOLOGICAL PROFILE BASED SPECTRAL-SPATIAL CLASSIFICATION SCHEME FOR HYPERSPECTRAL IMAGERY PARALLEL IMPLEMENTATION OF MORPHOLOGICAL PROFILE BASED SPECTRAL-SPATIAL CLASSIFICATION SCHEME FOR HYPERSPECTRAL IMAGERY B. Kumar a, O. Dikshit b a Department of Computer Science & Information Technology,

More information

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016

CPSC 340: Machine Learning and Data Mining. Principal Component Analysis Fall 2016 CPSC 340: Machine Learning and Data Mining Principal Component Analysis Fall 2016 A2/Midterm: Admin Grades/solutions will be posted after class. Assignment 4: Posted, due November 14. Extra office hours:

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

Post-Classification Change Detection of High Resolution Satellite Images Using AdaBoost Classifier

Post-Classification Change Detection of High Resolution Satellite Images Using AdaBoost Classifier , pp.34-38 http://dx.doi.org/10.14257/astl.2015.117.08 Post-Classification Change Detection of High Resolution Satellite Images Using AdaBoost Classifier Dong-Min Woo 1 and Viet Dung Do 1 1 Department

More information

Spatial Information Based Image Classification Using Support Vector Machine

Spatial Information Based Image Classification Using Support Vector Machine Spatial Information Based Image Classification Using Support Vector Machine P.Jeevitha, Dr. P. Ganesh Kumar PG Scholar, Dept of IT, Regional Centre of Anna University, Coimbatore, India. Assistant Professor,

More information

Constrained Manifold Learning for Hyperspectral Imagery Visualization

Constrained Manifold Learning for Hyperspectral Imagery Visualization IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING 1 Constrained Manifold Learning for Hyperspectral Imagery Visualization Danping Liao, Yuntao Qian Member, IEEE, and Yuan

More information

Semi-supervised Data Representation via Affinity Graph Learning

Semi-supervised Data Representation via Affinity Graph Learning 1 Semi-supervised Data Representation via Affinity Graph Learning Weiya Ren 1 1 College of Information System and Management, National University of Defense Technology, Changsha, Hunan, P.R China, 410073

More information

Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed

Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed Murinto #1, Nur Rochmah DPA #2 # Department of Informatics Engineering, Faculty of Industrial Technology,

More information

Band Selection for Hyperspectral Image Classification Using Mutual Information

Band Selection for Hyperspectral Image Classification Using Mutual Information 1 Band Selection for Hyperspectral Image Classification Using Mutual Information Baofeng Guo, Steve R. Gunn, R. I. Damper Senior Member, IEEE and J. D. B. Nelson Abstract Spectral band selection is a fundamental

More information

A Comparative Study of Conventional and Neural Network Classification of Multispectral Data

A Comparative Study of Conventional and Neural Network Classification of Multispectral Data A Comparative Study of Conventional and Neural Network Classification of Multispectral Data B.Solaiman & M.C.Mouchot Ecole Nationale Supérieure des Télécommunications de Bretagne B.P. 832, 29285 BREST

More information

Dimension reduction for hyperspectral imaging using laplacian eigenmaps and randomized principal component analysis:midyear Report

Dimension reduction for hyperspectral imaging using laplacian eigenmaps and randomized principal component analysis:midyear Report Dimension reduction for hyperspectral imaging using laplacian eigenmaps and randomized principal component analysis:midyear Report Yiran Li yl534@math.umd.edu Advisor: Wojtek Czaja wojtek@math.umd.edu

More information

HYPERSPECTRAL remote sensing imagery provides

HYPERSPECTRAL remote sensing imagery provides IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING 1 An Unsupervised Spectral Matching Classifier Based on Artificial DNA Computing for Hyperspectral Remote Sensing Imagery Hongzan Jiao, Yanfei Zhong,

More information

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES

GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES GENDER CLASSIFICATION USING SUPPORT VECTOR MACHINES Ashwin Swaminathan ashwins@umd.edu ENEE633: Statistical and Neural Pattern Recognition Instructor : Prof. Rama Chellappa Project 2, Part (a) 1. INTRODUCTION

More information

ROBUST JOINT SPARSITY MODEL FOR HYPERSPECTRAL IMAGE CLASSIFICATION. Wuhan University, China

ROBUST JOINT SPARSITY MODEL FOR HYPERSPECTRAL IMAGE CLASSIFICATION. Wuhan University, China ROBUST JOINT SPARSITY MODEL FOR HYPERSPECTRAL IMAGE CLASSIFICATION Shaoguang Huang 1, Hongyan Zhang 2, Wenzhi Liao 1 and Aleksandra Pižurica 1 1 Department of Telecommunications and Information Processing,

More information