HYPERSPECTRAL sensors provide a rich source of
|
|
- Philippa Powers
- 6 years ago
- Views:
Transcription
1 Fast Hyperspectral Feature Reduction Using Piecewise Constant Function Approximations Are C. Jensen, Student member, IEEE and Anne Schistad Solberg, Member, IEEE Abstract The high number of spectral bands obtained from hyperspectral sensors, combined with the often limited ground truth, solicits some kind of feature reduction when attempting supervised classification. This paper demonstrates that an optimal constant function representation of hyperspectral signature curves in the mean square sense is capable of representing the data sufficiently to outperform, or match, other feature reduction methods like PCA, SFS and DBFE for classification purposes on all of the four hyperspectral datasets that we have tested. The simple averaging of spectral bands makes the resulting features directly interpretable in a physical sense. Using an efficient dynamic programming algorithm the proposed method can be considered fast. Index Terms Pattern classification, Remote sensing, Feature extraction I. INTRODUCTION HYPERSPECTRAL sensors provide a rich source of information to allow an accurate separation of land cover classes. Often several hundred spectral samples are acquired for every pixel. Unfortunately, the number of pixels available for training the classifiers is often severely limited, and in combination with the high number of spectral bands, the occurrence of the so-called Hughes phenomenon is almost inevitable. Furthermore, the spectral samples often exhibit high correlation adding a redundancy that may obscure the information important for classification. To alleviate these problems one can reduce the dimensionality by feature reduction, or try to regularize parameters by biasing them towards simpler and more stable estimates. This paper focuses on dimensionality reduction. Classical techniques for feature reduction in the pattern recognition literature can be applied by considering the spectral samples as features. This includes feature selection algorithms like sequential forward selection (SFS), forwardbackwards selection, floating search and the more recently proposed fast constrained search algorithm [1][2][3], and linear transformations like principal components transform (PCA), Fisher s canonical transform, decision boundary feature extraction (DBFE) and the discrete wavelet transform [1][4][5]. The linear transforms have a disadvantage over the selection methods in that direct interpretation of the resulting features is difficult, i.e., the new features are linear combinations of all the spectral bands. On the other hand, selecting features when the number of possible feature candidates is large and the number of training samples is limited, can lead to incorrect and unstable selections [6]. This work was supported by the Norwegian Research Council. A. C. Jensen and A. S. Solberg are with the University of Oslo. The method presented in this paper divides the spectral curves into contiguous regions by piecewise constant function approximations. The extracted constants are then used as new features. The assumption is that the continuous curves that the features are samples of can be approximated sufficiently by such functions to allow subsequent classification. By using piecewise constants instead of higher order polynomials, the number of resulting features is minimized and the features are simple averages of contiguous spectral bands allowing straightforward interpretation. This gives a compromise between the generality of linear transforms and the interpretability of the feature selection techniques. Averages of contiguous spectral bands is also the end result in the top-down generalized local discriminant bases (TD- GLDB) procedure in [7]. Their algorithm recursively partitions the spectral curve into two sets of bands and replaces each final set of bands by their mean value. The criterion for selecting the partitioning breakpoint is either the performance on a validation set or using an estimate of the differences in class probability densities. The same paper also proposes a more computationally intensive strategy where the simple averaging to create features are replaced by Fisher s canonical transform, and using bottom-up search strategies. Both methods are relying on a pairwise classification framework. Another approach based on contiguous band averages is that of [8], which uses adaptations of selection methods to extract non-overlapping spectral regions. Instead of iteratively optimizing a classification criterion as in the latter approach, the proposed method focuses solely on minimizing the representation error when finding the breakpoints between the regions to average. Replacing the piecewise constants by linear functions and minimizing the representational error for the class mean curves using a greedy recursive method have been applied in [9]. In Section II it is shown that this method, with the addition of finding globally optimal breakpoints, can be seen as a special case of the proposed optimization framework when considering piecewise linear instead of piecewise constant segments. This paper shows that the fairly simple and direct approach proposed outperforms, or matches, methods like SFS, PCA, DBFE and TD-GLDB on all of the four hyperspectral data sets that we have tested. II. OPTIMAL PIECEWISE CONSTANT REPRESENTATION The goal is to partition the hyperspectral signatures into a fixed number of contiguous intervals with constant intensities
2 minimizing the mean square representation error. Let S ij be the jth feature in the ith pixel of a dataset with a total of N pixels with M features. S = {S ij 1 i N, 1 j M} is thus the collection of all spectral curves (the entire training dataset) available regardless of class membership. We seek a set of K breakpoints, P = {p 1, p 2,.., p K }, which define the contiguous intervals, I k = [p k, p k+1 ). Note that the K breakpoints are equal for all the classes. In our model each interval, for every pixel, is represented by a constant, µ ik R. The square representation error of the model is thus H = K k=1 i=1 N (S ij µ ik ) 2. (1) If the breakpoints, P, are given, one ends up with a simple square function w.r.t. the constants, µ ik, so the minimizer of (1) w.r.t. µ ik is given by, letting I k denote the number of elements in the k-th interval, µ ik = 1 S ij, (2) I k i.e., the mean value of each pixel s interval between breakpoints. What is left to be determined to minimize (1) is thus the locations of the breakpoints. These breakpoints can be found using dynamic programming. The algorithm presented here is based on [10], with the extension of allowing multiple spectral curves. A more detailed exposition of the dynamic programming can be found in the mentioned reference. The following gives the minimum of (1) for all K. Define B(K, m) = min H m (3) P,θ, P =K where H m means (1) where only the first m features are used, P = K means the use of K breakpoints, and θ is the collection of constants. For M m K > 1 we have, since (3) is decreasing with increasing K, the recurrence relation where B(K, m) = min K 1 r m 1 D [r,m] = N i=1 ( B(K 1, r) + D[r+1,m] ) min µ (4) m (S ij µ) 2. (5) j=r In short, we end up with Algorithm 1. Algorithm 1 Finding the minimum of (1) for all K. 1. Set B(1, m) = D [1,m], 1 m M. 2. Determine B(K, m), 1 < K m M, from (4), keeping track of rk,m giving its minimum. 3. Determine recursively breakpoints from the minimizers, r, for B(K, M), B(K 1, rk,m ),.... After finding the breakpoints, the constants, µ ik, from (2) are applied as features in the classification process. An example of a piecewise constant function approximation with K = 20 is shown in Figure 1. Although the complexity of the algorithm is O(NM 3 ), N being the number of pixels and M Fig. 1. Example of piecewise constant function approximation from the KSC dataset. Number of segments (K+1) is 21. the number spectral bands, the inner loop of the algorithm is fast. Examples of execution times can be found in Section IV. The procedure can be perceived as finding an orthogonal linear transform minimizing the square representational error, with the additional constraint that each row in the resulting matrix satisfies the criterion of representing the mean of a single curve segment. Thus, subsequently, the extracted features are obtained by simple matrix multiplication. A. Possible extensions of the approach To balance possibly unequal numbers of samples in the different classes, and allow a priori class weighting, (1) can be rewritten H = K N k=1 i=1 p(i) (S ij µ ik ) 2 (6) n(i) where p(i) and n(i) are the a priori likelihood and the number of pixels of the class of the ith pixel, respectively. After a similar alteration on (5), Algorithm 1 is applicable directly. Replacing the simple piecewise constant function model in (1) with higher order polynomials is straightforward, but by using, e.g. a linear model, the number of features for a given set of breakpoints is doubled. This increase in features would most often require subsequent feature selection when applied in a classification setting. However, the method still produces a linear transform, although not orthogonal. B. Summarizing the approach To summarize the approach, training consists of determining the K breakpoints and the corresponding segment means minimizing the square representation error. The extracted parameters are then used in a regular classifier, in our case a Gaussian Maximum Likelihood classifier.
3 A. The Datasets III. EXPERIMENTS To give a thorough analysis of the performance of the proposed method, it has been applied on four hyperspectral datasets containing widely different types of data, and with dimensions ranging from 81 to 176 bands. The first dataset, ROSIS, is from an airborne sensor, contains forest type data, is divided into three classes, has 81 spectral bands and has a pixel size of 5.6m. The second dataset, DC [11], is from an airborne sensor, contains urban type data, is divided into five classes, and has 150 bands. The third set, KSC [12], is from an airborne sensor (AVIRIS), contains vegetation type data, divided into 13 classes, has 176 spectral bands and has 18m pixels. The last dataset, BOTSWANA [12], is from a scene taken by the Hyperion sensor aboard the EO-1 satellite, contains vegetation type data, is divided into 14 classes, has 145 bands, and has a 30m pixel size. The average number of training pixels per class is 700, 600, 196 and 115 for the respective data sets. B. Methods To evaluate the performance of the different methods we applied the standard approach of separating the available ground-truthed data into roughly equally sized regions for training and testing, and report performance on the test data. As far as possible, we have made sure that all regions for training are spatially disjoint from the regions for testing, to avoid training on neighboring pixels that are correlated with test data. In the case of the two datasets with sufficient numbers of pixels, 5 repeated experiments were designed by sampling randomly equally sized sets for each class from the one half of the available data marked for training the classifier. The proposed method is compared with principal components transform (PCA), sequential forward selection (SFS) using the sum of the estimated Mahalanobis distances as feature evaluation function, decision boundary feature extraction (DBFE) using leave-one-out covariance matrix estimate [13], and top-down generalized local discriminant bases (TD- GLDB)[7] using their suggested log-odds probabilities criterion. Simple majority voting was used in the TD-GLDB pairwise framework, and results using the implementation in [14] is reported. The PCA is fast, widely used, and is based on signal representation error, while the DBFE is more elaborate with the inclusion of class separability in the optimization process. The TD-GLDB is based on averaging of adjacent bands, and thus have similarities to the proposed method, even though it requires a pairwise classification setting. More sophisticated methods for feature selection, e.g. forward-backward selection and floating selection [1], have been omitted because of their excessive execution times, and thereby their inappropriateness to be compared with the other faster methods. The extracted features are used in a Gaussian Maximum Likelihood (GML) classifier. The number of features, e.g. the number of principal components in PCA or the number of contiguous intervals in the proposed method, were chosen using standard 10-fold crossvalidation on each of the sampled training sets. The average performance and stability reported is based on the test data in these 5 experiments. Due to lack of knowledge about the true priors of the ground truth classes, the average performance over all classes is reported in the experiments. This is the case for all the data sets. Experiments using the proposed method on data where the mean value for each feature has been set to zero have also been conducted. This was done to segregate the general structure of the spectral curve with the variance found in the data. IV. RESULTS Table I shows that the overall misclassification rate when the optimal number of features is chosen is lower using the proposed method on three out of four datasets. The sequential forward selection (SFS) method has a slightly better performance on the DC dataset, but the proposed method still outperforms the PCA, DBFE and TD-GLDB approach. Except for the TD-GLDB, the misclassification rates of the methods can, in a natural way, be visualized as functions with regards to the number of features used. Figures 2 and 3 are representative examples showing the performance as such functions. On the BOTSWANA dataset, the proposed method and PCA have quite equal performance, but the proposed method has a slightly sharper drop in misclassification rate in the first part of the graph and a lower misclassification increase as the number of features increase beyond the optimal point. The KSC results shown in Figure 3 show a distinctly lower misclassification rate for the proposed method for large intervals of feature numbers. When the mean curve of the data has been subtracted, the proposed method has a small drop in performance in three out of four datasets, but is still comparable or better than the other methods. The variances of the misclassification rate, i.e., the instabilities of the classifiers, are lower for the proposed method, with an increase when subtracting the mean. Even though the estimated placement of the breakpoints in the proposed method are quite stable across the different trials for each dataset, the chosen number of regions vary, as is seen in Table II. Note that although the chosen number of regions varies with experiments, the error rate remains fairly stable. Algorithm runtimes are shown in Table III. The time it takes to find the possible features using the proposed method is merely a fraction of the time it takes to do the SFSs, and considerably lower than to do the DBFEs, even with the nonoptimized Matlab version of the proposed method used in this study. The execution times for the pairwise strategy of TD-GLDB increase drastically when applied to the datasets with many classes. Of course, the PCA needs only to solve an eigenvalue problem, and takes about a second. V. DISCUSSION It is apparent from the results that the spectral grouping and averaging in the proposed method are capable of representing the data and capture the essential variation within the spectral curve. When the mean of the data is subtracted, the general form of the spectral curves is removed and only the signal
4 TABLE I TEST RESULTS WITH THE OPTIMAL NUMBER OF FEATURES CHOSEN BY CROSSVALIDATION FOR THE FOUR DATASETS USING PCA, SFS, DBFE AND THE PROPOSED METHOD. MEAN MISCLASSIFICATION RATE AND ONE STANDARD DEVIATION IN PERCENT. PCA ± ± SFS ± ± DBFE ± ± TD-GLDB ± ± Proposed ± ± Proposed(µ=0) ± ± TABLE II THE MEAN NUMBER OF FEATURES AND ONE STANDARD DEVIATION FOR THE DIFFERENT FEATURE REDUCTION APPROACHES. PCA 17.2 ± ± SFS 19.6 ± ± DBFE 36.2 ± ± TD-GLDB a 6.9± ± Proposed 20.0 ± ± Proposed(µ=0) 21.2 ± ± a The means of the pairwise classifiers are reported. TABLE III EXECUTION TIMES IN SECONDS TO FIND ALL FEATURE SETS RANGING FROM ONE TO FULL DIMENSION. SFS DBFE TD-GLDB a Proposed a TD-GLDB finds final features directly. Fig. 2. Misclassification rate for different numbers of features using PCA, SFS and the proposed method on the BOTSWANA dataset. Crossvalidation estimated 12, 14, 16, and 20 features for the proposed method, PCA, SFS and DBFE, respectively. Fig. 3. Misclassification rate for different numbers of features using PCA, SFS and the proposed method on the KSC dataset. Crossvalidation estimated 21, 14, 18, and 11 features for the proposed method, PCA, SFS and DBFE, respectively. variance is used to determine the locations of the breakpoints for the constant function intervals. In this case, the results are generally the same, but with a little increase in misclassification error. The reason for the decrease in performance is probably that the variance dominates the process of selecting the breakpoints anyway, and that the general form of the spectral curve helps choose more general spectral curve-fitting intervals. The increase in classification stability aids the latter explanation. Consequently, in an actual application of the proposed method, the removal of the mean is generally not to be recommended. The TD-GLDB, with its pairwise classification framework, gives some classes improved discrimination, even though the overall misclassification rate is significantly higher. Also, it might be that a more advanced voting scheme would yield a slightly improved classification rate when using TD-GLDB. However, the (C 1)C/2 time complexity increase for a C- class problem using its pairwise strategy renders the method unfit for fast feature reduction when the classes are numerous. The low mean number of selected features using TD-GLDB as seen in Table II can be explained by the generally lower complexity of two-class problems. The simple spectral averaging in the proposed method lets the new features have a direct interpretation as opposed to the linear feature combinations resulting from PCA and DBFE. This, together with the computational efficiency and classification strength of the proposed method, makes it a natural candidate when analyzing hyperspectral images. The algorithm presented in Section II can easily be extended to use higher order polynomials, but that would preclude the simple interpretation, and increase the number, of the resulting features. Another obvious improvement of the proposed method would be to do a feature selection on the resulting features. Pilot studies indicate that a slight increase in performance can be achieved this way, but including it would conceal the simplicity, and dampen the computational efficiency, of the proposed approach.
5 VI. CONCLUSION This paper has demonstrated that an optimal piecewise constant function representation of hyperspectral signature curves in the mean square sense is capable of representing the data sufficiently to outperform, or match, other feature reduction methods like PCA, SFS, DBFE and TD-GLDB for classification purposes. The simple averaging of spectral bands makes the resulting features directly interpretable in a physical sense. Using an efficient dynamic programming algorithm the proposed method can be considered fast, making it a natural candidate when analyzing hyperspectral images. ACKNOWLEDGEMENTS We are grateful to the CESBIO institute in Toulouse, France for providing the ROSIS (Fontainebleau) dataset. REFERENCES [1] T. Hastie, R. Tibshirani, and J. Friedman, The elements of statistical learning: data mining, inference and prediction. Springer, [2] S. B. Serpico, M. D Inca, F. Melgani, and G. Moser, Comparison of feature reduction techniques for classification of hyperspectral remote sensing data, in Proc. SPIE, Image and Signal Processing for Remote Sensing VIII, S. B. Serpico, Ed., vol. 4885, 2003, pp [3] S. B. Serpico and L. Bruzzone, A new search algorithm for feature selection in hyperspectral remote sensing images, IEEE Trans. Geosci. Remote Sensing, vol. 39, no. 7, pp , July [4] C. Lee and D. A. Landgrebe, Feature extraction based on decision boundaries, IEEE Trans. Pattern Anal. Mach. Intell., vol. 15, no. 4, pp , [5] L. M. Bruce, C. H. Koger, and J. Li, Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction, IEEE Trans. Geosci. Remote Sensing, vol. 40, no. 10, pp , October [6] H. Schulerud and F. Albregtsen, Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces, Computer Methods and Programs in Biomedicine, vol. 73, no. 2, pp , [7] S. Kumar, J. Ghosh, and M. Crawford, Best-bases feature extraction algorithms for classification of hyperspectral data, IEEE Trans. Geosci. Remote Sensing, vol. 39, no. 7, pp , July [8] S. B. Serpico, M. D Inca, and G. Moser, Design of spectral channels for hyperspectral image classification, in Proceedings of the 2004 International Geoscience and Remote Sensing Symposium (IGARSS 04), vol. 2, 2004, pp [9] A. Henneguelle, J. Ghosh, and M. M. Crawford, Polyline feature extraction for land cover classification using hyperspectral data. in Proceedings of the 1st Indian International Conference on Artificial Intelligence, 2003, pp [10] G. Winkler and V. Liebscher, Smoothers for discontinuous signals, Journal of Nonparametric Statistics, vol. 14, pp , [11] D. A. Landgrebe, Signal Theory Methods in Multispectral Remote Sensing. Wiley-Interscience, [12] J. Ham, Y. Chen, M. M. Crawford, and J. Ghosh, Investigation of the random forest framework for classification of hyperspectral data, IEEE Trans. Geosci. Remote Sensing, vol. 43, no. 3, pp , March [13] J. Hoffbeck and D. Landgrebe, Covariance matrix estimation and classification with limited training data, IEEE Trans. Pattern Anal. Machine Intell., vol. 18, pp , July [14] P.Paclik, S.Verzakov, and R.P.W.Duin, Hypertools 2.0: The toolbox for spectral image analysis, 2005.
A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2
A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA Naoto Yokoya 1 and Akira Iwasaki 1 Graduate Student, Department of Aeronautics and Astronautics, The University of
More informationData: a collection of numbers or facts that require further processing before they are meaningful
Digital Image Classification Data vs. Information Data: a collection of numbers or facts that require further processing before they are meaningful Information: Derived knowledge from raw data. Something
More informationINF 4300 Classification III Anne Solberg The agenda today:
INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15
More informationCOSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor
COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality
More informationClassification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University
Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate
More informationRemote Sensing & Photogrammetry W4. Beata Hejmanowska Building C4, room 212, phone:
Remote Sensing & Photogrammetry W4 Beata Hejmanowska Building C4, room 212, phone: +4812 617 22 72 605 061 510 galia@agh.edu.pl 1 General procedures in image classification Conventional multispectral classification
More informationUndirected Graphical Models. Raul Queiroz Feitosa
Undirected Graphical Models Raul Queiroz Feitosa Pros and Cons Advantages of UGMs over DGMs UGMs are more natural for some domains (e.g. context-dependent entities) Discriminative UGMs (CRF) are better
More informationFeature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani
Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate
More informationPrincipal Component Image Interpretation A Logical and Statistical Approach
Principal Component Image Interpretation A Logical and Statistical Approach Md Shahid Latif M.Tech Student, Department of Remote Sensing, Birla Institute of Technology, Mesra Ranchi, Jharkhand-835215 Abstract
More informationUnsupervised Change Detection in Remote-Sensing Images using Modified Self-Organizing Feature Map Neural Network
Unsupervised Change Detection in Remote-Sensing Images using Modified Self-Organizing Feature Map Neural Network Swarnajyoti Patra, Susmita Ghosh Department of Computer Science and Engineering Jadavpur
More informationRANDOM FORESTS OF BINARY HIERARCHICAL CLASSIFIERS FOR ANALYSIS OF HYPERSPECTRAL DATA
RANDOM FORESTS OF BINARY HIERARCHICAL CLASSIFIERS FOR ANALYSIS OF HYPERSPECTRAL DATA Melba M. Crawford 1*, JiSoo Ham 1, Yangchi Chen 1, Joydeep Ghosh 2 1 Center for Space Research, 3925 W. Braker Lane,
More informationClassification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University
Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate
More informationA Robust Band Compression Technique for Hyperspectral Image Classification
A Robust Band Compression Technique for Hyperspectral Image Classification Qazi Sami ul Haq,Lixin Shi,Linmi Tao,Shiqiang Yang Key Laboratory of Pervasive Computing, Ministry of Education Department of
More informationLearning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009
Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer
More informationUnderstanding Clustering Supervising the unsupervised
Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data
More informationFast Sample Generation with Variational Bayesian for Limited Data Hyperspectral Image Classification
Fast Sample Generation with Variational Bayesian for Limited Data Hyperspectral Image Classification July 26, 2018 AmirAbbas Davari, Hasan Can Özkan, Andreas Maier, Christian Riess Pattern Recognition
More informationTemporal Modeling and Missing Data Estimation for MODIS Vegetation data
Temporal Modeling and Missing Data Estimation for MODIS Vegetation data Rie Honda 1 Introduction The Moderate Resolution Imaging Spectroradiometer (MODIS) is the primary instrument on board NASA s Earth
More informationMany are called, but few are chosen. Feature selection and error estimation in high dimensional spaces
Computer Methods and Programs in Biomedicine (2004) 73, 91 99 Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces Helene Schulerud a, *, Fritz Albregtsen
More informationSpectral Classification
Spectral Classification Spectral Classification Supervised versus Unsupervised Classification n Unsupervised Classes are determined by the computer. Also referred to as clustering n Supervised Classes
More informationGRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION
GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION Nasehe Jamshidpour a, Saeid Homayouni b, Abdolreza Safari a a Dept. of Geomatics Engineering, College of Engineering,
More informationTextural Features for Hyperspectral Pixel Classification
Textural Features for Hyperspectral Pixel Classification Olga Rajadell, Pedro García-Sevilla, and Filiberto Pla Depto. Lenguajes y Sistemas Informáticos Jaume I University, Campus Riu Sec s/n 12071 Castellón,
More informationONE OF THE challenging problems in processing highdimensional
182 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 36, NO. 1, JANUARY 1998 A Fast Two-Stage Classification Method for High-Dimensional Remote Sensing Data Te-Ming Tu, Chin-Hsing Chen, Jiunn-Lin
More informationBig Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1
Big Data Methods Chapter 5: Machine learning Big Data Methods, Chapter 5, Slide 1 5.1 Introduction to machine learning What is machine learning? Concerned with the study and development of algorithms that
More informationThe Elimination of Correlation Errors in PIV Processing
9 th International Symposium on Applications of Laser Techniques to Fluid Mechanics, Lisbon, Portugal, July, 1998 The Elimination of Correlation Errors in PIV Processing Douglas P. Hart Massachusetts Institute
More informationIntroduction to machine learning, pattern recognition and statistical data modelling Coryn Bailer-Jones
Introduction to machine learning, pattern recognition and statistical data modelling Coryn Bailer-Jones What is machine learning? Data interpretation describing relationship between predictors and responses
More informationNeural Network based textural labeling of images in multimedia applications
Neural Network based textural labeling of images in multimedia applications S.A. Karkanis +, G.D. Magoulas +, and D.A. Karras ++ + University of Athens, Dept. of Informatics, Typa Build., Panepistimiopolis,
More informationSupervised Classification in High Dimensional Space: Geometrical, Statistical and Asymptotical Properties of Multivariate Data 1
Supervised Classification in High Dimensional Space: Geometrical, Statistical and Asymptotical Properties of Multivariate Data 1 Luis Jimenez and David Landgrebe 2 Dept. Of ECE, PO Box 5000 School of Elect.
More informationMULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER
MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER A.Shabbir 1, 2 and G.Verdoolaege 1, 3 1 Department of Applied Physics, Ghent University, B-9000 Ghent, Belgium 2 Max Planck Institute
More informationIntroduction to digital image classification
Introduction to digital image classification Dr. Norman Kerle, Wan Bakx MSc a.o. INTERNATIONAL INSTITUTE FOR GEO-INFORMATION SCIENCE AND EARTH OBSERVATION Purpose of lecture Main lecture topics Review
More informationFeature Selection for Image Retrieval and Object Recognition
Feature Selection for Image Retrieval and Object Recognition Nuno Vasconcelos et al. Statistical Visual Computing Lab ECE, UCSD Presented by Dashan Gao Scalable Discriminant Feature Selection for Image
More informationHyperspectral Image Segmentation using Homogeneous Area Limiting and Shortest Path Algorithm
Hyperspectral Image Segmentation using Homogeneous Area Limiting and Shortest Path Algorithm Fatemeh Hajiani Department of Electrical Engineering, College of Engineering, Khormuj Branch, Islamic Azad University,
More informationModel Assessment and Selection. Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer
Model Assessment and Selection Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Model Training data Testing data Model Testing error rate Training error
More informationRobust Kernel Methods in Clustering and Dimensionality Reduction Problems
Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust
More informationA Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2
A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation Kwanyong Lee 1 and Hyeyoung Park 2 1. Department of Computer Science, Korea National Open
More informationSegmentation of Images
Segmentation of Images SEGMENTATION If an image has been preprocessed appropriately to remove noise and artifacts, segmentation is often the key step in interpreting the image. Image segmentation is a
More informationImproving Image Segmentation Quality Via Graph Theory
International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,
More informationCLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS
CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of
More informationThree Dimensional Texture Computation of Gray Level Co-occurrence Tensor in Hyperspectral Image Cubes
Three Dimensional Texture Computation of Gray Level Co-occurrence Tensor in Hyperspectral Image Cubes Jhe-Syuan Lai 1 and Fuan Tsai 2 Center for Space and Remote Sensing Research and Department of Civil
More informationFusion of pixel based and object based features for classification of urban hyperspectral remote sensing data
Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data Wenzhi liao a, *, Frieke Van Coillie b, Flore Devriendt b, Sidharta Gautama a, Aleksandra Pizurica
More informationLecture 11: Classification
Lecture 11: Classification 1 2009-04-28 Patrik Malm Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University 2 Reading instructions Chapters for this lecture 12.1 12.2 in
More informationRandom projection for non-gaussian mixture models
Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,
More informationFast Anomaly Detection Algorithms For Hyperspectral Images
Vol. Issue 9, September - 05 Fast Anomaly Detection Algorithms For Hyperspectral Images J. Zhou Google, Inc. ountain View, California, USA C. Kwan Signal Processing, Inc. Rockville, aryland, USA chiman.kwan@signalpro.net
More informationMulti-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification
Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification Selim Aksoy and H. Gökhan Akçay Bilkent University Department of Computer Engineering Bilkent, 06800, Ankara, Turkey
More informationUnsupervised Feature Selection for Sparse Data
Unsupervised Feature Selection for Sparse Data Artur Ferreira 1,3 Mário Figueiredo 2,3 1- Instituto Superior de Engenharia de Lisboa, Lisboa, PORTUGAL 2- Instituto Superior Técnico, Lisboa, PORTUGAL 3-
More informationBagging for One-Class Learning
Bagging for One-Class Learning David Kamm December 13, 2008 1 Introduction Consider the following outlier detection problem: suppose you are given an unlabeled data set and make the assumptions that one
More informationModel-based segmentation and recognition from range data
Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This
More informationPattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition
Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant
More informationMachine Learning with MATLAB --classification
Machine Learning with MATLAB --classification Stanley Liang, PhD York University Classification the definition In machine learning and statistics, classification is the problem of identifying to which
More informationSpatial Information Based Image Classification Using Support Vector Machine
Spatial Information Based Image Classification Using Support Vector Machine P.Jeevitha, Dr. P. Ganesh Kumar PG Scholar, Dept of IT, Regional Centre of Anna University, Coimbatore, India. Assistant Professor,
More informationFuzzy Entropy based feature selection for classification of hyperspectral data
Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering NIT Kurukshetra, 136119 mpce_pal@yahoo.co.uk Abstract: This paper proposes to use
More informationContextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation
IGARSS-2011 Vancouver, Canada, July 24-29, 29, 2011 Contextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation Gabriele Moser Sebastiano
More informationExtension of training set using mean shift procedure for aerospace images classification
Extension of training set using mean shift procedure for aerospace images classification Yuriy N. Sinyavskiy 1,*, Pavel V. Melnikov 1, and Igor A. Pestunov 1 1 Institute of Computational Technologies of
More informationApplication of nonparametric Bayesian classifier to remote sensing data. Institute of Parallel Processing, Bulgarian Academy of Sciences
Application of nonparametric Bayesian classifier to remote sensing data Nina Jeliazkova, nina@acad.bg, +359 2 979 6606 Stela Ruseva, stela@acad.bg, +359 2 979 6606 Kiril Boyanov, boyanov@acad.bg Institute
More informationHyperspectral Remote Sensing
Hyperspectral Remote Sensing Multi-spectral: Several comparatively wide spectral bands Hyperspectral: Many (could be hundreds) very narrow spectral bands GEOG 4110/5100 30 AVIRIS: Airborne Visible/Infrared
More informationDigital Image Classification Geography 4354 Remote Sensing
Digital Image Classification Geography 4354 Remote Sensing Lab 11 Dr. James Campbell December 10, 2001 Group #4 Mark Dougherty Paul Bartholomew Akisha Williams Dave Trible Seth McCoy Table of Contents:
More informationThis paper describes an analytical approach to the parametric analysis of target/decoy
Parametric analysis of target/decoy performance1 John P. Kerekes Lincoln Laboratory, Massachusetts Institute of Technology 244 Wood Street Lexington, Massachusetts 02173 ABSTRACT As infrared sensing technology
More informationNCC 2009, January 16-18, IIT Guwahati 267
NCC 2009, January 6-8, IIT Guwahati 267 Unsupervised texture segmentation based on Hadamard transform Tathagata Ray, Pranab Kumar Dutta Department Of Electrical Engineering Indian Institute of Technology
More informationHYPERSPECTRAL REMOTE SENSING
HYPERSPECTRAL REMOTE SENSING By Samuel Rosario Overview The Electromagnetic Spectrum Radiation Types MSI vs HIS Sensors Applications Image Analysis Software Feature Extraction Information Extraction 1
More informationRemote Sensed Image Classification based on Spatial and Spectral Features using SVM
RESEARCH ARTICLE OPEN ACCESS Remote Sensed Image Classification based on Spatial and Spectral Features using SVM Mary Jasmine. E PG Scholar Department of Computer Science and Engineering, University College
More informationFace Recognition using Eigenfaces SMAI Course Project
Face Recognition using Eigenfaces SMAI Course Project Satarupa Guha IIIT Hyderabad 201307566 satarupa.guha@research.iiit.ac.in Ayushi Dalmia IIIT Hyderabad 201307565 ayushi.dalmia@research.iiit.ac.in Abstract
More informationSupport Vector Selection and Adaptation and Its Application in Remote Sensing
Support Vector Selection and Adaptation and Its Application in Remote Sensing Gülşen Taşkın Kaya Computational Science and Engineering Istanbul Technical University Istanbul, Turkey gtaskink@purdue.edu
More informationLinear Methods for Regression and Shrinkage Methods
Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors
More informationSpatially variant dimensionality reduction for the visualization of multi/hyperspectral images
Author manuscript, published in "International Conference on Image Analysis and Recognition, Burnaby : Canada (2011)" DOI : 10.1007/978-3-642-21593-3_38 Spatially variant dimensionality reduction for the
More informationNoise-based Feature Perturbation as a Selection Method for Microarray Data
Noise-based Feature Perturbation as a Selection Method for Microarray Data Li Chen 1, Dmitry B. Goldgof 1, Lawrence O. Hall 1, and Steven A. Eschrich 2 1 Department of Computer Science and Engineering
More informationUnsupervised Change Detection in Optical Satellite Images using Binary Descriptor
Unsupervised Change Detection in Optical Satellite Images using Binary Descriptor Neha Gupta, Gargi V. Pillai, Samit Ari Department of Electronics and Communication Engineering, National Institute of Technology,
More informationNonlinear data separation and fusion for multispectral image classification
Nonlinear data separation and fusion for multispectral image classification Hela Elmannai #*1, Mohamed Anis Loghmari #2, Mohamed Saber Naceur #3 # Laboratoire de Teledetection et Systeme d informations
More informationChapter 2 Basic Structure of High-Dimensional Spaces
Chapter 2 Basic Structure of High-Dimensional Spaces Data is naturally represented geometrically by associating each record with a point in the space spanned by the attributes. This idea, although simple,
More informationRandom Forest A. Fornaser
Random Forest A. Fornaser alberto.fornaser@unitn.it Sources Lecture 15: decision trees, information theory and random forests, Dr. Richard E. Turner Trees and Random Forests, Adele Cutler, Utah State University
More informationHYPERSPECTRAL imaging has received considerable. A Signal-Decomposed and Interference-Annihilated Approach to Hyperspectral Target Detection
892 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 42, NO.4, APRIL 2004 A Signal-Decomposed and Interference-Annihilated Approach to Hyperspectral Target Detection Qian Du, Member, IEEE, and
More informationPERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION
20th European Signal Processing Conference EUSIPCO 2012) Bucharest, Romania, August 27-31, 2012 PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION Mauricio Lara 1 and Bernard Mulgrew
More informationEquation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation.
Equation to LaTeX Abhinav Rastogi, Sevy Harris {arastogi,sharris5}@stanford.edu I. Introduction Copying equations from a pdf file to a LaTeX document can be time consuming because there is no easy way
More informationCS6716 Pattern Recognition
CS6716 Pattern Recognition Prototype Methods Aaron Bobick School of Interactive Computing Administrivia Problem 2b was extended to March 25. Done? PS3 will be out this real soon (tonight) due April 10.
More informationClassification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed
Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed Murinto #1, Nur Rochmah DPA #2 # Department of Informatics Engineering, Faculty of Industrial Technology,
More informationIMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING
SECOND EDITION IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING ith Algorithms for ENVI/IDL Morton J. Canty с*' Q\ CRC Press Taylor &. Francis Group Boca Raton London New York CRC
More informationDEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION
DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION S.Dhanalakshmi #1 #PG Scholar, Department of Computer Science, Dr.Sivanthi Aditanar college of Engineering, Tiruchendur
More informationStatistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte
Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,
More informationAN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION
AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION WILLIAM ROBSON SCHWARTZ University of Maryland, Department of Computer Science College Park, MD, USA, 20742-327, schwartz@cs.umd.edu RICARDO
More informationClassification of Hyperspectral Data over Urban. Areas Using Directional Morphological Profiles and. Semi-supervised Feature Extraction
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL.X, NO.X, Y 1 Classification of Hyperspectral Data over Urban Areas Using Directional Morphological Profiles and Semi-supervised
More informationDecision trees. Decision trees are useful to a large degree because of their simplicity and interpretability
Decision trees A decision tree is a method for classification/regression that aims to ask a few relatively simple questions about an input and then predicts the associated output Decision trees are useful
More informationPattern Recognition & Classification
CEE 6150: Digital Image Processing 1 Classification Supervised parallelpiped minimum distance maximum likelihood (Bayes Rule) > non-parametric > parametric Unsupervised (clustering) K-Means ISODATA support
More informationCS 195-5: Machine Learning Problem Set 5
CS 195-5: Machine Learning Problem Set 5 Douglas Lanman dlanman@brown.edu 26 November 26 1 Clustering and Vector Quantization Problem 1 Part 1: In this problem we will apply Vector Quantization (VQ) to
More informationDimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image
Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image Murinto 1, Nur Rochmah Dyah PA 2 1,2 Department of Informatics Engineering
More informationAdaptive Metric Nearest Neighbor Classification
Adaptive Metric Nearest Neighbor Classification Carlotta Domeniconi Jing Peng Dimitrios Gunopulos Computer Science Department Computer Science Department Computer Science Department University of California
More informationHYPERSPECTRAL imagery (HSI) records hundreds of
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 11, NO. 1, JANUARY 2014 173 Classification Based on 3-D DWT and Decision Fusion for Hyperspectral Image Analysis Zhen Ye, Student Member, IEEE, Saurabh
More informationQUALITATIVE MODELING FOR MAGNETIZATION CURVE
Journal of Marine Science and Technology, Vol. 8, No. 2, pp. 65-70 (2000) 65 QUALITATIVE MODELING FOR MAGNETIZATION CURVE Pei-Hwa Huang and Yu-Shuo Chang Keywords: Magnetization curve, Qualitative modeling,
More informationEstimating the Natural Number of Classes on Hierarchically Clustered Multi-spectral Images
Estimating the Natural Number of Classes on Hierarchically Clustered Multi-spectral Images André R.S. Marçal and Janete S. Borges Faculdade de Ciências, Universidade do Porto, DMA, Rua do Campo Alegre,
More informationRobust Face Recognition via Sparse Representation
Robust Face Recognition via Sparse Representation Panqu Wang Department of Electrical and Computer Engineering University of California, San Diego La Jolla, CA 92092 pawang@ucsd.edu Can Xu Department of
More informationClassification of Hyperspectral Breast Images for Cancer Detection. Sander Parawira December 4, 2009
1 Introduction Classification of Hyperspectral Breast Images for Cancer Detection Sander Parawira December 4, 2009 parawira@stanford.edu In 2009 approximately one out of eight women has breast cancer.
More informationLab 9. Julia Janicki. Introduction
Lab 9 Julia Janicki Introduction My goal for this project is to map a general land cover in the area of Alexandria in Egypt using supervised classification, specifically the Maximum Likelihood and Support
More informationIEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE 2014 2147 Automatic Framework for Spectral Spatial Classification Based on Supervised Feature Extraction
More informationSupervised vs. Unsupervised Learning
Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now
More information38050 Povo Trento (Italy), Via Sommarive 14
UNIVERSITY OF TRENTO DEPARTMENT OF INFORMATION AND COMMUNICATION TECHNOLOGY 38050 Povo Trento (Italy), Via Sommarive 14 http://www.dit.unitn.it A NEW SEARCH ALGORITHM FOR FEATURE SELECTION IN HYPERSPECTRAL
More informationCS 229 Midterm Review
CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask
More informationINTELLIGENT TARGET DETECTION IN HYPERSPECTRAL IMAGERY
INTELLIGENT TARGET DETECTION IN HYPERSPECTRAL IMAGERY Ayanna Howard, Curtis Padgett, Kenneth Brown Jet Propulsion Laboratory, California Institute of Technology 4800 Oak Grove Drive, Pasadena, CA 91 109-8099
More informationAudio-Visual Speech Activity Detection
Institut für Technische Informatik und Kommunikationsnetze Semester Thesis at the Department of Information Technology and Electrical Engineering Audio-Visual Speech Activity Detection Salome Mannale Advisors:
More informationFOOTPRINTS EXTRACTION
Building Footprints Extraction of Dense Residential Areas from LiDAR data KyoHyouk Kim and Jie Shan Purdue University School of Civil Engineering 550 Stadium Mall Drive West Lafayette, IN 47907, USA {kim458,
More informationSATELLITE IMAGE CLASSIFICATION USING WAVELET TRANSFORM
International Journal of Electronics and Communication Engineering & Technology (IJECET) ISSN 0976 6464(Print), ISSN 0976 6472(Online) Volume 1, Number 1, Sep - Oct (2010), pp. 117-124 IAEME, http://www.iaeme.com/ijecet.html
More informationNetwork Traffic Measurements and Analysis
DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:
More informationMultidirectional 2DPCA Based Face Recognition System
Multidirectional 2DPCA Based Face Recognition System Shilpi Soni 1, Raj Kumar Sahu 2 1 M.E. Scholar, Department of E&Tc Engg, CSIT, Durg 2 Associate Professor, Department of E&Tc Engg, CSIT, Durg Email:
More informationIntroduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering
Introduction to Pattern Recognition Part II Selim Aksoy Bilkent University Department of Computer Engineering saksoy@cs.bilkent.edu.tr RETINA Pattern Recognition Tutorial, Summer 2005 Overview Statistical
More information