HYPERSPECTRAL sensors provide a rich source of

Size: px
Start display at page:

Download "HYPERSPECTRAL sensors provide a rich source of"

Transcription

1 Fast Hyperspectral Feature Reduction Using Piecewise Constant Function Approximations Are C. Jensen, Student member, IEEE and Anne Schistad Solberg, Member, IEEE Abstract The high number of spectral bands obtained from hyperspectral sensors, combined with the often limited ground truth, solicits some kind of feature reduction when attempting supervised classification. This paper demonstrates that an optimal constant function representation of hyperspectral signature curves in the mean square sense is capable of representing the data sufficiently to outperform, or match, other feature reduction methods like PCA, SFS and DBFE for classification purposes on all of the four hyperspectral datasets that we have tested. The simple averaging of spectral bands makes the resulting features directly interpretable in a physical sense. Using an efficient dynamic programming algorithm the proposed method can be considered fast. Index Terms Pattern classification, Remote sensing, Feature extraction I. INTRODUCTION HYPERSPECTRAL sensors provide a rich source of information to allow an accurate separation of land cover classes. Often several hundred spectral samples are acquired for every pixel. Unfortunately, the number of pixels available for training the classifiers is often severely limited, and in combination with the high number of spectral bands, the occurrence of the so-called Hughes phenomenon is almost inevitable. Furthermore, the spectral samples often exhibit high correlation adding a redundancy that may obscure the information important for classification. To alleviate these problems one can reduce the dimensionality by feature reduction, or try to regularize parameters by biasing them towards simpler and more stable estimates. This paper focuses on dimensionality reduction. Classical techniques for feature reduction in the pattern recognition literature can be applied by considering the spectral samples as features. This includes feature selection algorithms like sequential forward selection (SFS), forwardbackwards selection, floating search and the more recently proposed fast constrained search algorithm [1][2][3], and linear transformations like principal components transform (PCA), Fisher s canonical transform, decision boundary feature extraction (DBFE) and the discrete wavelet transform [1][4][5]. The linear transforms have a disadvantage over the selection methods in that direct interpretation of the resulting features is difficult, i.e., the new features are linear combinations of all the spectral bands. On the other hand, selecting features when the number of possible feature candidates is large and the number of training samples is limited, can lead to incorrect and unstable selections [6]. This work was supported by the Norwegian Research Council. A. C. Jensen and A. S. Solberg are with the University of Oslo. The method presented in this paper divides the spectral curves into contiguous regions by piecewise constant function approximations. The extracted constants are then used as new features. The assumption is that the continuous curves that the features are samples of can be approximated sufficiently by such functions to allow subsequent classification. By using piecewise constants instead of higher order polynomials, the number of resulting features is minimized and the features are simple averages of contiguous spectral bands allowing straightforward interpretation. This gives a compromise between the generality of linear transforms and the interpretability of the feature selection techniques. Averages of contiguous spectral bands is also the end result in the top-down generalized local discriminant bases (TD- GLDB) procedure in [7]. Their algorithm recursively partitions the spectral curve into two sets of bands and replaces each final set of bands by their mean value. The criterion for selecting the partitioning breakpoint is either the performance on a validation set or using an estimate of the differences in class probability densities. The same paper also proposes a more computationally intensive strategy where the simple averaging to create features are replaced by Fisher s canonical transform, and using bottom-up search strategies. Both methods are relying on a pairwise classification framework. Another approach based on contiguous band averages is that of [8], which uses adaptations of selection methods to extract non-overlapping spectral regions. Instead of iteratively optimizing a classification criterion as in the latter approach, the proposed method focuses solely on minimizing the representation error when finding the breakpoints between the regions to average. Replacing the piecewise constants by linear functions and minimizing the representational error for the class mean curves using a greedy recursive method have been applied in [9]. In Section II it is shown that this method, with the addition of finding globally optimal breakpoints, can be seen as a special case of the proposed optimization framework when considering piecewise linear instead of piecewise constant segments. This paper shows that the fairly simple and direct approach proposed outperforms, or matches, methods like SFS, PCA, DBFE and TD-GLDB on all of the four hyperspectral data sets that we have tested. II. OPTIMAL PIECEWISE CONSTANT REPRESENTATION The goal is to partition the hyperspectral signatures into a fixed number of contiguous intervals with constant intensities

2 minimizing the mean square representation error. Let S ij be the jth feature in the ith pixel of a dataset with a total of N pixels with M features. S = {S ij 1 i N, 1 j M} is thus the collection of all spectral curves (the entire training dataset) available regardless of class membership. We seek a set of K breakpoints, P = {p 1, p 2,.., p K }, which define the contiguous intervals, I k = [p k, p k+1 ). Note that the K breakpoints are equal for all the classes. In our model each interval, for every pixel, is represented by a constant, µ ik R. The square representation error of the model is thus H = K k=1 i=1 N (S ij µ ik ) 2. (1) If the breakpoints, P, are given, one ends up with a simple square function w.r.t. the constants, µ ik, so the minimizer of (1) w.r.t. µ ik is given by, letting I k denote the number of elements in the k-th interval, µ ik = 1 S ij, (2) I k i.e., the mean value of each pixel s interval between breakpoints. What is left to be determined to minimize (1) is thus the locations of the breakpoints. These breakpoints can be found using dynamic programming. The algorithm presented here is based on [10], with the extension of allowing multiple spectral curves. A more detailed exposition of the dynamic programming can be found in the mentioned reference. The following gives the minimum of (1) for all K. Define B(K, m) = min H m (3) P,θ, P =K where H m means (1) where only the first m features are used, P = K means the use of K breakpoints, and θ is the collection of constants. For M m K > 1 we have, since (3) is decreasing with increasing K, the recurrence relation where B(K, m) = min K 1 r m 1 D [r,m] = N i=1 ( B(K 1, r) + D[r+1,m] ) min µ (4) m (S ij µ) 2. (5) j=r In short, we end up with Algorithm 1. Algorithm 1 Finding the minimum of (1) for all K. 1. Set B(1, m) = D [1,m], 1 m M. 2. Determine B(K, m), 1 < K m M, from (4), keeping track of rk,m giving its minimum. 3. Determine recursively breakpoints from the minimizers, r, for B(K, M), B(K 1, rk,m ),.... After finding the breakpoints, the constants, µ ik, from (2) are applied as features in the classification process. An example of a piecewise constant function approximation with K = 20 is shown in Figure 1. Although the complexity of the algorithm is O(NM 3 ), N being the number of pixels and M Fig. 1. Example of piecewise constant function approximation from the KSC dataset. Number of segments (K+1) is 21. the number spectral bands, the inner loop of the algorithm is fast. Examples of execution times can be found in Section IV. The procedure can be perceived as finding an orthogonal linear transform minimizing the square representational error, with the additional constraint that each row in the resulting matrix satisfies the criterion of representing the mean of a single curve segment. Thus, subsequently, the extracted features are obtained by simple matrix multiplication. A. Possible extensions of the approach To balance possibly unequal numbers of samples in the different classes, and allow a priori class weighting, (1) can be rewritten H = K N k=1 i=1 p(i) (S ij µ ik ) 2 (6) n(i) where p(i) and n(i) are the a priori likelihood and the number of pixels of the class of the ith pixel, respectively. After a similar alteration on (5), Algorithm 1 is applicable directly. Replacing the simple piecewise constant function model in (1) with higher order polynomials is straightforward, but by using, e.g. a linear model, the number of features for a given set of breakpoints is doubled. This increase in features would most often require subsequent feature selection when applied in a classification setting. However, the method still produces a linear transform, although not orthogonal. B. Summarizing the approach To summarize the approach, training consists of determining the K breakpoints and the corresponding segment means minimizing the square representation error. The extracted parameters are then used in a regular classifier, in our case a Gaussian Maximum Likelihood classifier.

3 A. The Datasets III. EXPERIMENTS To give a thorough analysis of the performance of the proposed method, it has been applied on four hyperspectral datasets containing widely different types of data, and with dimensions ranging from 81 to 176 bands. The first dataset, ROSIS, is from an airborne sensor, contains forest type data, is divided into three classes, has 81 spectral bands and has a pixel size of 5.6m. The second dataset, DC [11], is from an airborne sensor, contains urban type data, is divided into five classes, and has 150 bands. The third set, KSC [12], is from an airborne sensor (AVIRIS), contains vegetation type data, divided into 13 classes, has 176 spectral bands and has 18m pixels. The last dataset, BOTSWANA [12], is from a scene taken by the Hyperion sensor aboard the EO-1 satellite, contains vegetation type data, is divided into 14 classes, has 145 bands, and has a 30m pixel size. The average number of training pixels per class is 700, 600, 196 and 115 for the respective data sets. B. Methods To evaluate the performance of the different methods we applied the standard approach of separating the available ground-truthed data into roughly equally sized regions for training and testing, and report performance on the test data. As far as possible, we have made sure that all regions for training are spatially disjoint from the regions for testing, to avoid training on neighboring pixels that are correlated with test data. In the case of the two datasets with sufficient numbers of pixels, 5 repeated experiments were designed by sampling randomly equally sized sets for each class from the one half of the available data marked for training the classifier. The proposed method is compared with principal components transform (PCA), sequential forward selection (SFS) using the sum of the estimated Mahalanobis distances as feature evaluation function, decision boundary feature extraction (DBFE) using leave-one-out covariance matrix estimate [13], and top-down generalized local discriminant bases (TD- GLDB)[7] using their suggested log-odds probabilities criterion. Simple majority voting was used in the TD-GLDB pairwise framework, and results using the implementation in [14] is reported. The PCA is fast, widely used, and is based on signal representation error, while the DBFE is more elaborate with the inclusion of class separability in the optimization process. The TD-GLDB is based on averaging of adjacent bands, and thus have similarities to the proposed method, even though it requires a pairwise classification setting. More sophisticated methods for feature selection, e.g. forward-backward selection and floating selection [1], have been omitted because of their excessive execution times, and thereby their inappropriateness to be compared with the other faster methods. The extracted features are used in a Gaussian Maximum Likelihood (GML) classifier. The number of features, e.g. the number of principal components in PCA or the number of contiguous intervals in the proposed method, were chosen using standard 10-fold crossvalidation on each of the sampled training sets. The average performance and stability reported is based on the test data in these 5 experiments. Due to lack of knowledge about the true priors of the ground truth classes, the average performance over all classes is reported in the experiments. This is the case for all the data sets. Experiments using the proposed method on data where the mean value for each feature has been set to zero have also been conducted. This was done to segregate the general structure of the spectral curve with the variance found in the data. IV. RESULTS Table I shows that the overall misclassification rate when the optimal number of features is chosen is lower using the proposed method on three out of four datasets. The sequential forward selection (SFS) method has a slightly better performance on the DC dataset, but the proposed method still outperforms the PCA, DBFE and TD-GLDB approach. Except for the TD-GLDB, the misclassification rates of the methods can, in a natural way, be visualized as functions with regards to the number of features used. Figures 2 and 3 are representative examples showing the performance as such functions. On the BOTSWANA dataset, the proposed method and PCA have quite equal performance, but the proposed method has a slightly sharper drop in misclassification rate in the first part of the graph and a lower misclassification increase as the number of features increase beyond the optimal point. The KSC results shown in Figure 3 show a distinctly lower misclassification rate for the proposed method for large intervals of feature numbers. When the mean curve of the data has been subtracted, the proposed method has a small drop in performance in three out of four datasets, but is still comparable or better than the other methods. The variances of the misclassification rate, i.e., the instabilities of the classifiers, are lower for the proposed method, with an increase when subtracting the mean. Even though the estimated placement of the breakpoints in the proposed method are quite stable across the different trials for each dataset, the chosen number of regions vary, as is seen in Table II. Note that although the chosen number of regions varies with experiments, the error rate remains fairly stable. Algorithm runtimes are shown in Table III. The time it takes to find the possible features using the proposed method is merely a fraction of the time it takes to do the SFSs, and considerably lower than to do the DBFEs, even with the nonoptimized Matlab version of the proposed method used in this study. The execution times for the pairwise strategy of TD-GLDB increase drastically when applied to the datasets with many classes. Of course, the PCA needs only to solve an eigenvalue problem, and takes about a second. V. DISCUSSION It is apparent from the results that the spectral grouping and averaging in the proposed method are capable of representing the data and capture the essential variation within the spectral curve. When the mean of the data is subtracted, the general form of the spectral curves is removed and only the signal

4 TABLE I TEST RESULTS WITH THE OPTIMAL NUMBER OF FEATURES CHOSEN BY CROSSVALIDATION FOR THE FOUR DATASETS USING PCA, SFS, DBFE AND THE PROPOSED METHOD. MEAN MISCLASSIFICATION RATE AND ONE STANDARD DEVIATION IN PERCENT. PCA ± ± SFS ± ± DBFE ± ± TD-GLDB ± ± Proposed ± ± Proposed(µ=0) ± ± TABLE II THE MEAN NUMBER OF FEATURES AND ONE STANDARD DEVIATION FOR THE DIFFERENT FEATURE REDUCTION APPROACHES. PCA 17.2 ± ± SFS 19.6 ± ± DBFE 36.2 ± ± TD-GLDB a 6.9± ± Proposed 20.0 ± ± Proposed(µ=0) 21.2 ± ± a The means of the pairwise classifiers are reported. TABLE III EXECUTION TIMES IN SECONDS TO FIND ALL FEATURE SETS RANGING FROM ONE TO FULL DIMENSION. SFS DBFE TD-GLDB a Proposed a TD-GLDB finds final features directly. Fig. 2. Misclassification rate for different numbers of features using PCA, SFS and the proposed method on the BOTSWANA dataset. Crossvalidation estimated 12, 14, 16, and 20 features for the proposed method, PCA, SFS and DBFE, respectively. Fig. 3. Misclassification rate for different numbers of features using PCA, SFS and the proposed method on the KSC dataset. Crossvalidation estimated 21, 14, 18, and 11 features for the proposed method, PCA, SFS and DBFE, respectively. variance is used to determine the locations of the breakpoints for the constant function intervals. In this case, the results are generally the same, but with a little increase in misclassification error. The reason for the decrease in performance is probably that the variance dominates the process of selecting the breakpoints anyway, and that the general form of the spectral curve helps choose more general spectral curve-fitting intervals. The increase in classification stability aids the latter explanation. Consequently, in an actual application of the proposed method, the removal of the mean is generally not to be recommended. The TD-GLDB, with its pairwise classification framework, gives some classes improved discrimination, even though the overall misclassification rate is significantly higher. Also, it might be that a more advanced voting scheme would yield a slightly improved classification rate when using TD-GLDB. However, the (C 1)C/2 time complexity increase for a C- class problem using its pairwise strategy renders the method unfit for fast feature reduction when the classes are numerous. The low mean number of selected features using TD-GLDB as seen in Table II can be explained by the generally lower complexity of two-class problems. The simple spectral averaging in the proposed method lets the new features have a direct interpretation as opposed to the linear feature combinations resulting from PCA and DBFE. This, together with the computational efficiency and classification strength of the proposed method, makes it a natural candidate when analyzing hyperspectral images. The algorithm presented in Section II can easily be extended to use higher order polynomials, but that would preclude the simple interpretation, and increase the number, of the resulting features. Another obvious improvement of the proposed method would be to do a feature selection on the resulting features. Pilot studies indicate that a slight increase in performance can be achieved this way, but including it would conceal the simplicity, and dampen the computational efficiency, of the proposed approach.

5 VI. CONCLUSION This paper has demonstrated that an optimal piecewise constant function representation of hyperspectral signature curves in the mean square sense is capable of representing the data sufficiently to outperform, or match, other feature reduction methods like PCA, SFS, DBFE and TD-GLDB for classification purposes. The simple averaging of spectral bands makes the resulting features directly interpretable in a physical sense. Using an efficient dynamic programming algorithm the proposed method can be considered fast, making it a natural candidate when analyzing hyperspectral images. ACKNOWLEDGEMENTS We are grateful to the CESBIO institute in Toulouse, France for providing the ROSIS (Fontainebleau) dataset. REFERENCES [1] T. Hastie, R. Tibshirani, and J. Friedman, The elements of statistical learning: data mining, inference and prediction. Springer, [2] S. B. Serpico, M. D Inca, F. Melgani, and G. Moser, Comparison of feature reduction techniques for classification of hyperspectral remote sensing data, in Proc. SPIE, Image and Signal Processing for Remote Sensing VIII, S. B. Serpico, Ed., vol. 4885, 2003, pp [3] S. B. Serpico and L. Bruzzone, A new search algorithm for feature selection in hyperspectral remote sensing images, IEEE Trans. Geosci. Remote Sensing, vol. 39, no. 7, pp , July [4] C. Lee and D. A. Landgrebe, Feature extraction based on decision boundaries, IEEE Trans. Pattern Anal. Mach. Intell., vol. 15, no. 4, pp , [5] L. M. Bruce, C. H. Koger, and J. Li, Dimensionality reduction of hyperspectral data using discrete wavelet transform feature extraction, IEEE Trans. Geosci. Remote Sensing, vol. 40, no. 10, pp , October [6] H. Schulerud and F. Albregtsen, Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces, Computer Methods and Programs in Biomedicine, vol. 73, no. 2, pp , [7] S. Kumar, J. Ghosh, and M. Crawford, Best-bases feature extraction algorithms for classification of hyperspectral data, IEEE Trans. Geosci. Remote Sensing, vol. 39, no. 7, pp , July [8] S. B. Serpico, M. D Inca, and G. Moser, Design of spectral channels for hyperspectral image classification, in Proceedings of the 2004 International Geoscience and Remote Sensing Symposium (IGARSS 04), vol. 2, 2004, pp [9] A. Henneguelle, J. Ghosh, and M. M. Crawford, Polyline feature extraction for land cover classification using hyperspectral data. in Proceedings of the 1st Indian International Conference on Artificial Intelligence, 2003, pp [10] G. Winkler and V. Liebscher, Smoothers for discontinuous signals, Journal of Nonparametric Statistics, vol. 14, pp , [11] D. A. Landgrebe, Signal Theory Methods in Multispectral Remote Sensing. Wiley-Interscience, [12] J. Ham, Y. Chen, M. M. Crawford, and J. Ghosh, Investigation of the random forest framework for classification of hyperspectral data, IEEE Trans. Geosci. Remote Sensing, vol. 43, no. 3, pp , March [13] J. Hoffbeck and D. Landgrebe, Covariance matrix estimation and classification with limited training data, IEEE Trans. Pattern Anal. Machine Intell., vol. 18, pp , July [14] P.Paclik, S.Verzakov, and R.P.W.Duin, Hypertools 2.0: The toolbox for spectral image analysis, 2005.

A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2

A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA. Naoto Yokoya 1 and Akira Iwasaki 2 A MAXIMUM NOISE FRACTION TRANSFORM BASED ON A SENSOR NOISE MODEL FOR HYPERSPECTRAL DATA Naoto Yokoya 1 and Akira Iwasaki 1 Graduate Student, Department of Aeronautics and Astronautics, The University of

More information

Data: a collection of numbers or facts that require further processing before they are meaningful

Data: a collection of numbers or facts that require further processing before they are meaningful Digital Image Classification Data vs. Information Data: a collection of numbers or facts that require further processing before they are meaningful Information: Derived knowledge from raw data. Something

More information

INF 4300 Classification III Anne Solberg The agenda today:

INF 4300 Classification III Anne Solberg The agenda today: INF 4300 Classification III Anne Solberg 28.10.15 The agenda today: More on estimating classifier accuracy Curse of dimensionality and simple feature selection knn-classification K-means clustering 28.10.15

More information

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor COSC160: Detection and Classification Jeremy Bolton, PhD Assistant Teaching Professor Outline I. Problem I. Strategies II. Features for training III. Using spatial information? IV. Reducing dimensionality

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

Remote Sensing & Photogrammetry W4. Beata Hejmanowska Building C4, room 212, phone:

Remote Sensing & Photogrammetry W4. Beata Hejmanowska Building C4, room 212, phone: Remote Sensing & Photogrammetry W4 Beata Hejmanowska Building C4, room 212, phone: +4812 617 22 72 605 061 510 galia@agh.edu.pl 1 General procedures in image classification Conventional multispectral classification

More information

Undirected Graphical Models. Raul Queiroz Feitosa

Undirected Graphical Models. Raul Queiroz Feitosa Undirected Graphical Models Raul Queiroz Feitosa Pros and Cons Advantages of UGMs over DGMs UGMs are more natural for some domains (e.g. context-dependent entities) Discriminative UGMs (CRF) are better

More information

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani

Feature Selection. CE-725: Statistical Pattern Recognition Sharif University of Technology Spring Soleymani Feature Selection CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Dimensionality reduction Feature selection vs. feature extraction Filter univariate

More information

Principal Component Image Interpretation A Logical and Statistical Approach

Principal Component Image Interpretation A Logical and Statistical Approach Principal Component Image Interpretation A Logical and Statistical Approach Md Shahid Latif M.Tech Student, Department of Remote Sensing, Birla Institute of Technology, Mesra Ranchi, Jharkhand-835215 Abstract

More information

Unsupervised Change Detection in Remote-Sensing Images using Modified Self-Organizing Feature Map Neural Network

Unsupervised Change Detection in Remote-Sensing Images using Modified Self-Organizing Feature Map Neural Network Unsupervised Change Detection in Remote-Sensing Images using Modified Self-Organizing Feature Map Neural Network Swarnajyoti Patra, Susmita Ghosh Department of Computer Science and Engineering Jadavpur

More information

RANDOM FORESTS OF BINARY HIERARCHICAL CLASSIFIERS FOR ANALYSIS OF HYPERSPECTRAL DATA

RANDOM FORESTS OF BINARY HIERARCHICAL CLASSIFIERS FOR ANALYSIS OF HYPERSPECTRAL DATA RANDOM FORESTS OF BINARY HIERARCHICAL CLASSIFIERS FOR ANALYSIS OF HYPERSPECTRAL DATA Melba M. Crawford 1*, JiSoo Ham 1, Yangchi Chen 1, Joydeep Ghosh 2 1 Center for Space Research, 3925 W. Braker Lane,

More information

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Classification Vladimir Curic Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University Outline An overview on classification Basics of classification How to choose appropriate

More information

A Robust Band Compression Technique for Hyperspectral Image Classification

A Robust Band Compression Technique for Hyperspectral Image Classification A Robust Band Compression Technique for Hyperspectral Image Classification Qazi Sami ul Haq,Lixin Shi,Linmi Tao,Shiqiang Yang Key Laboratory of Pervasive Computing, Ministry of Education Department of

More information

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009

Learning and Inferring Depth from Monocular Images. Jiyan Pan April 1, 2009 Learning and Inferring Depth from Monocular Images Jiyan Pan April 1, 2009 Traditional ways of inferring depth Binocular disparity Structure from motion Defocus Given a single monocular image, how to infer

More information

Understanding Clustering Supervising the unsupervised

Understanding Clustering Supervising the unsupervised Understanding Clustering Supervising the unsupervised Janu Verma IBM T.J. Watson Research Center, New York http://jverma.github.io/ jverma@us.ibm.com @januverma Clustering Grouping together similar data

More information

Fast Sample Generation with Variational Bayesian for Limited Data Hyperspectral Image Classification

Fast Sample Generation with Variational Bayesian for Limited Data Hyperspectral Image Classification Fast Sample Generation with Variational Bayesian for Limited Data Hyperspectral Image Classification July 26, 2018 AmirAbbas Davari, Hasan Can Özkan, Andreas Maier, Christian Riess Pattern Recognition

More information

Temporal Modeling and Missing Data Estimation for MODIS Vegetation data

Temporal Modeling and Missing Data Estimation for MODIS Vegetation data Temporal Modeling and Missing Data Estimation for MODIS Vegetation data Rie Honda 1 Introduction The Moderate Resolution Imaging Spectroradiometer (MODIS) is the primary instrument on board NASA s Earth

More information

Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces

Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces Computer Methods and Programs in Biomedicine (2004) 73, 91 99 Many are called, but few are chosen. Feature selection and error estimation in high dimensional spaces Helene Schulerud a, *, Fritz Albregtsen

More information

Spectral Classification

Spectral Classification Spectral Classification Spectral Classification Supervised versus Unsupervised Classification n Unsupervised Classes are determined by the computer. Also referred to as clustering n Supervised Classes

More information

GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION

GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION GRAPH-BASED SEMI-SUPERVISED HYPERSPECTRAL IMAGE CLASSIFICATION USING SPATIAL INFORMATION Nasehe Jamshidpour a, Saeid Homayouni b, Abdolreza Safari a a Dept. of Geomatics Engineering, College of Engineering,

More information

Textural Features for Hyperspectral Pixel Classification

Textural Features for Hyperspectral Pixel Classification Textural Features for Hyperspectral Pixel Classification Olga Rajadell, Pedro García-Sevilla, and Filiberto Pla Depto. Lenguajes y Sistemas Informáticos Jaume I University, Campus Riu Sec s/n 12071 Castellón,

More information

ONE OF THE challenging problems in processing highdimensional

ONE OF THE challenging problems in processing highdimensional 182 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 36, NO. 1, JANUARY 1998 A Fast Two-Stage Classification Method for High-Dimensional Remote Sensing Data Te-Ming Tu, Chin-Hsing Chen, Jiunn-Lin

More information

Big Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1

Big Data Methods. Chapter 5: Machine learning. Big Data Methods, Chapter 5, Slide 1 Big Data Methods Chapter 5: Machine learning Big Data Methods, Chapter 5, Slide 1 5.1 Introduction to machine learning What is machine learning? Concerned with the study and development of algorithms that

More information

The Elimination of Correlation Errors in PIV Processing

The Elimination of Correlation Errors in PIV Processing 9 th International Symposium on Applications of Laser Techniques to Fluid Mechanics, Lisbon, Portugal, July, 1998 The Elimination of Correlation Errors in PIV Processing Douglas P. Hart Massachusetts Institute

More information

Introduction to machine learning, pattern recognition and statistical data modelling Coryn Bailer-Jones

Introduction to machine learning, pattern recognition and statistical data modelling Coryn Bailer-Jones Introduction to machine learning, pattern recognition and statistical data modelling Coryn Bailer-Jones What is machine learning? Data interpretation describing relationship between predictors and responses

More information

Neural Network based textural labeling of images in multimedia applications

Neural Network based textural labeling of images in multimedia applications Neural Network based textural labeling of images in multimedia applications S.A. Karkanis +, G.D. Magoulas +, and D.A. Karras ++ + University of Athens, Dept. of Informatics, Typa Build., Panepistimiopolis,

More information

Supervised Classification in High Dimensional Space: Geometrical, Statistical and Asymptotical Properties of Multivariate Data 1

Supervised Classification in High Dimensional Space: Geometrical, Statistical and Asymptotical Properties of Multivariate Data 1 Supervised Classification in High Dimensional Space: Geometrical, Statistical and Asymptotical Properties of Multivariate Data 1 Luis Jimenez and David Landgrebe 2 Dept. Of ECE, PO Box 5000 School of Elect.

More information

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER

MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER MULTIVARIATE TEXTURE DISCRIMINATION USING A PRINCIPAL GEODESIC CLASSIFIER A.Shabbir 1, 2 and G.Verdoolaege 1, 3 1 Department of Applied Physics, Ghent University, B-9000 Ghent, Belgium 2 Max Planck Institute

More information

Introduction to digital image classification

Introduction to digital image classification Introduction to digital image classification Dr. Norman Kerle, Wan Bakx MSc a.o. INTERNATIONAL INSTITUTE FOR GEO-INFORMATION SCIENCE AND EARTH OBSERVATION Purpose of lecture Main lecture topics Review

More information

Feature Selection for Image Retrieval and Object Recognition

Feature Selection for Image Retrieval and Object Recognition Feature Selection for Image Retrieval and Object Recognition Nuno Vasconcelos et al. Statistical Visual Computing Lab ECE, UCSD Presented by Dashan Gao Scalable Discriminant Feature Selection for Image

More information

Hyperspectral Image Segmentation using Homogeneous Area Limiting and Shortest Path Algorithm

Hyperspectral Image Segmentation using Homogeneous Area Limiting and Shortest Path Algorithm Hyperspectral Image Segmentation using Homogeneous Area Limiting and Shortest Path Algorithm Fatemeh Hajiani Department of Electrical Engineering, College of Engineering, Khormuj Branch, Islamic Azad University,

More information

Model Assessment and Selection. Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer

Model Assessment and Selection. Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer Model Assessment and Selection Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Model Training data Testing data Model Testing error rate Training error

More information

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems

Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Robust Kernel Methods in Clustering and Dimensionality Reduction Problems Jian Guo, Debadyuti Roy, Jing Wang University of Michigan, Department of Statistics Introduction In this report we propose robust

More information

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2

A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation. Kwanyong Lee 1 and Hyeyoung Park 2 A Distance-Based Classifier Using Dissimilarity Based on Class Conditional Probability and Within-Class Variation Kwanyong Lee 1 and Hyeyoung Park 2 1. Department of Computer Science, Korea National Open

More information

Segmentation of Images

Segmentation of Images Segmentation of Images SEGMENTATION If an image has been preprocessed appropriately to remove noise and artifacts, segmentation is often the key step in interpreting the image. Image segmentation is a

More information

Improving Image Segmentation Quality Via Graph Theory

Improving Image Segmentation Quality Via Graph Theory International Symposium on Computers & Informatics (ISCI 05) Improving Image Segmentation Quality Via Graph Theory Xiangxiang Li, Songhao Zhu School of Automatic, Nanjing University of Post and Telecommunications,

More information

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS

CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS CHAPTER 4 CLASSIFICATION WITH RADIAL BASIS AND PROBABILISTIC NEURAL NETWORKS 4.1 Introduction Optical character recognition is one of

More information

Three Dimensional Texture Computation of Gray Level Co-occurrence Tensor in Hyperspectral Image Cubes

Three Dimensional Texture Computation of Gray Level Co-occurrence Tensor in Hyperspectral Image Cubes Three Dimensional Texture Computation of Gray Level Co-occurrence Tensor in Hyperspectral Image Cubes Jhe-Syuan Lai 1 and Fuan Tsai 2 Center for Space and Remote Sensing Research and Department of Civil

More information

Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data

Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data Fusion of pixel based and object based features for classification of urban hyperspectral remote sensing data Wenzhi liao a, *, Frieke Van Coillie b, Flore Devriendt b, Sidharta Gautama a, Aleksandra Pizurica

More information

Lecture 11: Classification

Lecture 11: Classification Lecture 11: Classification 1 2009-04-28 Patrik Malm Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University 2 Reading instructions Chapters for this lecture 12.1 12.2 in

More information

Random projection for non-gaussian mixture models

Random projection for non-gaussian mixture models Random projection for non-gaussian mixture models Győző Gidófalvi Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 92037 gyozo@cs.ucsd.edu Abstract Recently,

More information

Fast Anomaly Detection Algorithms For Hyperspectral Images

Fast Anomaly Detection Algorithms For Hyperspectral Images Vol. Issue 9, September - 05 Fast Anomaly Detection Algorithms For Hyperspectral Images J. Zhou Google, Inc. ountain View, California, USA C. Kwan Signal Processing, Inc. Rockville, aryland, USA chiman.kwan@signalpro.net

More information

Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification

Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification Multi-resolution Segmentation and Shape Analysis for Remote Sensing Image Classification Selim Aksoy and H. Gökhan Akçay Bilkent University Department of Computer Engineering Bilkent, 06800, Ankara, Turkey

More information

Unsupervised Feature Selection for Sparse Data

Unsupervised Feature Selection for Sparse Data Unsupervised Feature Selection for Sparse Data Artur Ferreira 1,3 Mário Figueiredo 2,3 1- Instituto Superior de Engenharia de Lisboa, Lisboa, PORTUGAL 2- Instituto Superior Técnico, Lisboa, PORTUGAL 3-

More information

Bagging for One-Class Learning

Bagging for One-Class Learning Bagging for One-Class Learning David Kamm December 13, 2008 1 Introduction Consider the following outlier detection problem: suppose you are given an unlabeled data set and make the assumptions that one

More information

Model-based segmentation and recognition from range data

Model-based segmentation and recognition from range data Model-based segmentation and recognition from range data Jan Boehm Institute for Photogrammetry Universität Stuttgart Germany Keywords: range image, segmentation, object recognition, CAD ABSTRACT This

More information

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition

Pattern Recognition. Kjell Elenius. Speech, Music and Hearing KTH. March 29, 2007 Speech recognition Pattern Recognition Kjell Elenius Speech, Music and Hearing KTH March 29, 2007 Speech recognition 2007 1 Ch 4. Pattern Recognition 1(3) Bayes Decision Theory Minimum-Error-Rate Decision Rules Discriminant

More information

Machine Learning with MATLAB --classification

Machine Learning with MATLAB --classification Machine Learning with MATLAB --classification Stanley Liang, PhD York University Classification the definition In machine learning and statistics, classification is the problem of identifying to which

More information

Spatial Information Based Image Classification Using Support Vector Machine

Spatial Information Based Image Classification Using Support Vector Machine Spatial Information Based Image Classification Using Support Vector Machine P.Jeevitha, Dr. P. Ganesh Kumar PG Scholar, Dept of IT, Regional Centre of Anna University, Coimbatore, India. Assistant Professor,

More information

Fuzzy Entropy based feature selection for classification of hyperspectral data

Fuzzy Entropy based feature selection for classification of hyperspectral data Fuzzy Entropy based feature selection for classification of hyperspectral data Mahesh Pal Department of Civil Engineering NIT Kurukshetra, 136119 mpce_pal@yahoo.co.uk Abstract: This paper proposes to use

More information

Contextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation

Contextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation IGARSS-2011 Vancouver, Canada, July 24-29, 29, 2011 Contextual High-Resolution Image Classification by Markovian Data Fusion, Adaptive Texture Extraction, and Multiscale Segmentation Gabriele Moser Sebastiano

More information

Extension of training set using mean shift procedure for aerospace images classification

Extension of training set using mean shift procedure for aerospace images classification Extension of training set using mean shift procedure for aerospace images classification Yuriy N. Sinyavskiy 1,*, Pavel V. Melnikov 1, and Igor A. Pestunov 1 1 Institute of Computational Technologies of

More information

Application of nonparametric Bayesian classifier to remote sensing data. Institute of Parallel Processing, Bulgarian Academy of Sciences

Application of nonparametric Bayesian classifier to remote sensing data. Institute of Parallel Processing, Bulgarian Academy of Sciences Application of nonparametric Bayesian classifier to remote sensing data Nina Jeliazkova, nina@acad.bg, +359 2 979 6606 Stela Ruseva, stela@acad.bg, +359 2 979 6606 Kiril Boyanov, boyanov@acad.bg Institute

More information

Hyperspectral Remote Sensing

Hyperspectral Remote Sensing Hyperspectral Remote Sensing Multi-spectral: Several comparatively wide spectral bands Hyperspectral: Many (could be hundreds) very narrow spectral bands GEOG 4110/5100 30 AVIRIS: Airborne Visible/Infrared

More information

Digital Image Classification Geography 4354 Remote Sensing

Digital Image Classification Geography 4354 Remote Sensing Digital Image Classification Geography 4354 Remote Sensing Lab 11 Dr. James Campbell December 10, 2001 Group #4 Mark Dougherty Paul Bartholomew Akisha Williams Dave Trible Seth McCoy Table of Contents:

More information

This paper describes an analytical approach to the parametric analysis of target/decoy

This paper describes an analytical approach to the parametric analysis of target/decoy Parametric analysis of target/decoy performance1 John P. Kerekes Lincoln Laboratory, Massachusetts Institute of Technology 244 Wood Street Lexington, Massachusetts 02173 ABSTRACT As infrared sensing technology

More information

NCC 2009, January 16-18, IIT Guwahati 267

NCC 2009, January 16-18, IIT Guwahati 267 NCC 2009, January 6-8, IIT Guwahati 267 Unsupervised texture segmentation based on Hadamard transform Tathagata Ray, Pranab Kumar Dutta Department Of Electrical Engineering Indian Institute of Technology

More information

HYPERSPECTRAL REMOTE SENSING

HYPERSPECTRAL REMOTE SENSING HYPERSPECTRAL REMOTE SENSING By Samuel Rosario Overview The Electromagnetic Spectrum Radiation Types MSI vs HIS Sensors Applications Image Analysis Software Feature Extraction Information Extraction 1

More information

Remote Sensed Image Classification based on Spatial and Spectral Features using SVM

Remote Sensed Image Classification based on Spatial and Spectral Features using SVM RESEARCH ARTICLE OPEN ACCESS Remote Sensed Image Classification based on Spatial and Spectral Features using SVM Mary Jasmine. E PG Scholar Department of Computer Science and Engineering, University College

More information

Face Recognition using Eigenfaces SMAI Course Project

Face Recognition using Eigenfaces SMAI Course Project Face Recognition using Eigenfaces SMAI Course Project Satarupa Guha IIIT Hyderabad 201307566 satarupa.guha@research.iiit.ac.in Ayushi Dalmia IIIT Hyderabad 201307565 ayushi.dalmia@research.iiit.ac.in Abstract

More information

Support Vector Selection and Adaptation and Its Application in Remote Sensing

Support Vector Selection and Adaptation and Its Application in Remote Sensing Support Vector Selection and Adaptation and Its Application in Remote Sensing Gülşen Taşkın Kaya Computational Science and Engineering Istanbul Technical University Istanbul, Turkey gtaskink@purdue.edu

More information

Linear Methods for Regression and Shrinkage Methods

Linear Methods for Regression and Shrinkage Methods Linear Methods for Regression and Shrinkage Methods Reference: The Elements of Statistical Learning, by T. Hastie, R. Tibshirani, J. Friedman, Springer 1 Linear Regression Models Least Squares Input vectors

More information

Spatially variant dimensionality reduction for the visualization of multi/hyperspectral images

Spatially variant dimensionality reduction for the visualization of multi/hyperspectral images Author manuscript, published in "International Conference on Image Analysis and Recognition, Burnaby : Canada (2011)" DOI : 10.1007/978-3-642-21593-3_38 Spatially variant dimensionality reduction for the

More information

Noise-based Feature Perturbation as a Selection Method for Microarray Data

Noise-based Feature Perturbation as a Selection Method for Microarray Data Noise-based Feature Perturbation as a Selection Method for Microarray Data Li Chen 1, Dmitry B. Goldgof 1, Lawrence O. Hall 1, and Steven A. Eschrich 2 1 Department of Computer Science and Engineering

More information

Unsupervised Change Detection in Optical Satellite Images using Binary Descriptor

Unsupervised Change Detection in Optical Satellite Images using Binary Descriptor Unsupervised Change Detection in Optical Satellite Images using Binary Descriptor Neha Gupta, Gargi V. Pillai, Samit Ari Department of Electronics and Communication Engineering, National Institute of Technology,

More information

Nonlinear data separation and fusion for multispectral image classification

Nonlinear data separation and fusion for multispectral image classification Nonlinear data separation and fusion for multispectral image classification Hela Elmannai #*1, Mohamed Anis Loghmari #2, Mohamed Saber Naceur #3 # Laboratoire de Teledetection et Systeme d informations

More information

Chapter 2 Basic Structure of High-Dimensional Spaces

Chapter 2 Basic Structure of High-Dimensional Spaces Chapter 2 Basic Structure of High-Dimensional Spaces Data is naturally represented geometrically by associating each record with a point in the space spanned by the attributes. This idea, although simple,

More information

Random Forest A. Fornaser

Random Forest A. Fornaser Random Forest A. Fornaser alberto.fornaser@unitn.it Sources Lecture 15: decision trees, information theory and random forests, Dr. Richard E. Turner Trees and Random Forests, Adele Cutler, Utah State University

More information

HYPERSPECTRAL imaging has received considerable. A Signal-Decomposed and Interference-Annihilated Approach to Hyperspectral Target Detection

HYPERSPECTRAL imaging has received considerable. A Signal-Decomposed and Interference-Annihilated Approach to Hyperspectral Target Detection 892 IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 42, NO.4, APRIL 2004 A Signal-Decomposed and Interference-Annihilated Approach to Hyperspectral Target Detection Qian Du, Member, IEEE, and

More information

PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION

PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION 20th European Signal Processing Conference EUSIPCO 2012) Bucharest, Romania, August 27-31, 2012 PERFORMANCE OF THE DISTRIBUTED KLT AND ITS APPROXIMATE IMPLEMENTATION Mauricio Lara 1 and Bernard Mulgrew

More information

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation.

Equation to LaTeX. Abhinav Rastogi, Sevy Harris. I. Introduction. Segmentation. Equation to LaTeX Abhinav Rastogi, Sevy Harris {arastogi,sharris5}@stanford.edu I. Introduction Copying equations from a pdf file to a LaTeX document can be time consuming because there is no easy way

More information

CS6716 Pattern Recognition

CS6716 Pattern Recognition CS6716 Pattern Recognition Prototype Methods Aaron Bobick School of Interactive Computing Administrivia Problem 2b was extended to March 25. Done? PS3 will be out this real soon (tonight) due April 10.

More information

Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed

Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed Classification of Hyper spectral Image Using Support Vector Machine and Marker-Controlled Watershed Murinto #1, Nur Rochmah DPA #2 # Department of Informatics Engineering, Faculty of Industrial Technology,

More information

IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING

IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING SECOND EDITION IMAGE ANALYSIS, CLASSIFICATION, and CHANGE DETECTION in REMOTE SENSING ith Algorithms for ENVI/IDL Morton J. Canty с*' Q\ CRC Press Taylor &. Francis Group Boca Raton London New York CRC

More information

DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION

DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION DEEP LEARNING TO DIVERSIFY BELIEF NETWORKS FOR REMOTE SENSING IMAGE CLASSIFICATION S.Dhanalakshmi #1 #PG Scholar, Department of Computer Science, Dr.Sivanthi Aditanar college of Engineering, Tiruchendur

More information

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte

Statistical Analysis of Metabolomics Data. Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Statistical Analysis of Metabolomics Data Xiuxia Du Department of Bioinformatics & Genomics University of North Carolina at Charlotte Outline Introduction Data pre-treatment 1. Normalization 2. Centering,

More information

AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION

AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION AN IMPROVED K-MEANS CLUSTERING ALGORITHM FOR IMAGE SEGMENTATION WILLIAM ROBSON SCHWARTZ University of Maryland, Department of Computer Science College Park, MD, USA, 20742-327, schwartz@cs.umd.edu RICARDO

More information

Classification of Hyperspectral Data over Urban. Areas Using Directional Morphological Profiles and. Semi-supervised Feature Extraction

Classification of Hyperspectral Data over Urban. Areas Using Directional Morphological Profiles and. Semi-supervised Feature Extraction IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL.X, NO.X, Y 1 Classification of Hyperspectral Data over Urban Areas Using Directional Morphological Profiles and Semi-supervised

More information

Decision trees. Decision trees are useful to a large degree because of their simplicity and interpretability

Decision trees. Decision trees are useful to a large degree because of their simplicity and interpretability Decision trees A decision tree is a method for classification/regression that aims to ask a few relatively simple questions about an input and then predicts the associated output Decision trees are useful

More information

Pattern Recognition & Classification

Pattern Recognition & Classification CEE 6150: Digital Image Processing 1 Classification Supervised parallelpiped minimum distance maximum likelihood (Bayes Rule) > non-parametric > parametric Unsupervised (clustering) K-Means ISODATA support

More information

CS 195-5: Machine Learning Problem Set 5

CS 195-5: Machine Learning Problem Set 5 CS 195-5: Machine Learning Problem Set 5 Douglas Lanman dlanman@brown.edu 26 November 26 1 Clustering and Vector Quantization Problem 1 Part 1: In this problem we will apply Vector Quantization (VQ) to

More information

Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image

Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image Dimensionality Reduction using Hybrid Support Vector Machine and Discriminant Independent Component Analysis for Hyperspectral Image Murinto 1, Nur Rochmah Dyah PA 2 1,2 Department of Informatics Engineering

More information

Adaptive Metric Nearest Neighbor Classification

Adaptive Metric Nearest Neighbor Classification Adaptive Metric Nearest Neighbor Classification Carlotta Domeniconi Jing Peng Dimitrios Gunopulos Computer Science Department Computer Science Department Computer Science Department University of California

More information

HYPERSPECTRAL imagery (HSI) records hundreds of

HYPERSPECTRAL imagery (HSI) records hundreds of IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, VOL. 11, NO. 1, JANUARY 2014 173 Classification Based on 3-D DWT and Decision Fusion for Hyperspectral Image Analysis Zhen Ye, Student Member, IEEE, Saurabh

More information

QUALITATIVE MODELING FOR MAGNETIZATION CURVE

QUALITATIVE MODELING FOR MAGNETIZATION CURVE Journal of Marine Science and Technology, Vol. 8, No. 2, pp. 65-70 (2000) 65 QUALITATIVE MODELING FOR MAGNETIZATION CURVE Pei-Hwa Huang and Yu-Shuo Chang Keywords: Magnetization curve, Qualitative modeling,

More information

Estimating the Natural Number of Classes on Hierarchically Clustered Multi-spectral Images

Estimating the Natural Number of Classes on Hierarchically Clustered Multi-spectral Images Estimating the Natural Number of Classes on Hierarchically Clustered Multi-spectral Images André R.S. Marçal and Janete S. Borges Faculdade de Ciências, Universidade do Porto, DMA, Rua do Campo Alegre,

More information

Robust Face Recognition via Sparse Representation

Robust Face Recognition via Sparse Representation Robust Face Recognition via Sparse Representation Panqu Wang Department of Electrical and Computer Engineering University of California, San Diego La Jolla, CA 92092 pawang@ucsd.edu Can Xu Department of

More information

Classification of Hyperspectral Breast Images for Cancer Detection. Sander Parawira December 4, 2009

Classification of Hyperspectral Breast Images for Cancer Detection. Sander Parawira December 4, 2009 1 Introduction Classification of Hyperspectral Breast Images for Cancer Detection Sander Parawira December 4, 2009 parawira@stanford.edu In 2009 approximately one out of eight women has breast cancer.

More information

Lab 9. Julia Janicki. Introduction

Lab 9. Julia Janicki. Introduction Lab 9 Julia Janicki Introduction My goal for this project is to map a general land cover in the area of Alexandria in Egypt using supervised classification, specifically the Maximum Likelihood and Support

More information

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE

IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, VOL. 7, NO. 6, JUNE 2014 2147 Automatic Framework for Spectral Spatial Classification Based on Supervised Feature Extraction

More information

Supervised vs. Unsupervised Learning

Supervised vs. Unsupervised Learning Clustering Supervised vs. Unsupervised Learning So far we have assumed that the training samples used to design the classifier were labeled by their class membership (supervised learning) We assume now

More information

38050 Povo Trento (Italy), Via Sommarive 14

38050 Povo Trento (Italy), Via Sommarive 14 UNIVERSITY OF TRENTO DEPARTMENT OF INFORMATION AND COMMUNICATION TECHNOLOGY 38050 Povo Trento (Italy), Via Sommarive 14 http://www.dit.unitn.it A NEW SEARCH ALGORITHM FOR FEATURE SELECTION IN HYPERSPECTRAL

More information

CS 229 Midterm Review

CS 229 Midterm Review CS 229 Midterm Review Course Staff Fall 2018 11/2/2018 Outline Today: SVMs Kernels Tree Ensembles EM Algorithm / Mixture Models [ Focus on building intuition, less so on solving specific problems. Ask

More information

INTELLIGENT TARGET DETECTION IN HYPERSPECTRAL IMAGERY

INTELLIGENT TARGET DETECTION IN HYPERSPECTRAL IMAGERY INTELLIGENT TARGET DETECTION IN HYPERSPECTRAL IMAGERY Ayanna Howard, Curtis Padgett, Kenneth Brown Jet Propulsion Laboratory, California Institute of Technology 4800 Oak Grove Drive, Pasadena, CA 91 109-8099

More information

Audio-Visual Speech Activity Detection

Audio-Visual Speech Activity Detection Institut für Technische Informatik und Kommunikationsnetze Semester Thesis at the Department of Information Technology and Electrical Engineering Audio-Visual Speech Activity Detection Salome Mannale Advisors:

More information

FOOTPRINTS EXTRACTION

FOOTPRINTS EXTRACTION Building Footprints Extraction of Dense Residential Areas from LiDAR data KyoHyouk Kim and Jie Shan Purdue University School of Civil Engineering 550 Stadium Mall Drive West Lafayette, IN 47907, USA {kim458,

More information

SATELLITE IMAGE CLASSIFICATION USING WAVELET TRANSFORM

SATELLITE IMAGE CLASSIFICATION USING WAVELET TRANSFORM International Journal of Electronics and Communication Engineering & Technology (IJECET) ISSN 0976 6464(Print), ISSN 0976 6472(Online) Volume 1, Number 1, Sep - Oct (2010), pp. 117-124 IAEME, http://www.iaeme.com/ijecet.html

More information

Network Traffic Measurements and Analysis

Network Traffic Measurements and Analysis DEIB - Politecnico di Milano Fall, 2017 Sources Hastie, Tibshirani, Friedman: The Elements of Statistical Learning James, Witten, Hastie, Tibshirani: An Introduction to Statistical Learning Andrew Ng:

More information

Multidirectional 2DPCA Based Face Recognition System

Multidirectional 2DPCA Based Face Recognition System Multidirectional 2DPCA Based Face Recognition System Shilpi Soni 1, Raj Kumar Sahu 2 1 M.E. Scholar, Department of E&Tc Engg, CSIT, Durg 2 Associate Professor, Department of E&Tc Engg, CSIT, Durg Email:

More information

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering

Introduction to Pattern Recognition Part II. Selim Aksoy Bilkent University Department of Computer Engineering Introduction to Pattern Recognition Part II Selim Aksoy Bilkent University Department of Computer Engineering saksoy@cs.bilkent.edu.tr RETINA Pattern Recognition Tutorial, Summer 2005 Overview Statistical

More information