Lab #4 Introduction to Image Processing II and Map Accuracy Assessment

Similar documents
Data: a collection of numbers or facts that require further processing before they are meaningful

Lab 9. Julia Janicki. Introduction

Chapter Seventeen: Classification of Remotely Sensed Imagery Introduction. Supervised Versus Unsupervised Classification

Classify Multi-Spectral Data Classify Geologic Terrains on Venus Apply Multi-Variate Statistics

DIGITAL IMAGE ANALYSIS. Image Classification: Object-based Classification

Digital Image Classification Geography 4354 Remote Sensing

Introduction to digital image classification

ENVI Classic Tutorial: A Quick Start to ENVI Classic

(Refer Slide Time: 0:51)

ENVI Tutorial: Introduction to ENVI

ENVI Classic Tutorial: Introduction to ENVI Classic 2

Raster Classification with ArcGIS Desktop. Rebecca Richman Andy Shoemaker

Attribute Accuracy. Quantitative accuracy refers to the level of bias in estimating the values assigned such as estimated values of ph in a soil map.

Classification (or thematic) accuracy assessment. Lecture 8 March 11, 2005

This is the general guide for landuse mapping using mid-resolution remote sensing data

Lab 5: Image Analysis with ArcGIS 10 Unsupervised Classification

Aardobservatie en Data-analyse Image processing

MultiSpec Tutorial. January 11, Program Concept and Introduction Notes by David Landgrebe and Larry Biehl. MultiSpec Programming by Larry Biehl

Geomatica II Course guide

Image Classification. RS Image Classification. Present by: Dr.Weerakaset Suanpaga

Land Cover Classification Techniques

Using ArcGIS for Landcover Classification. Presented by CORE GIS May 8, 2012

Classifying. Stuart Green Earthobservation.wordpress.com MERMS 12 - L4

v Importing Rasters SMS 11.2 Tutorial Requirements Raster Module Map Module Mesh Module Time minutes Prerequisites Overview Tutorial

Figure 1: Workflow of object-based classification

Files Used in This Tutorial. Background. Feature Extraction with Example-Based Classification Tutorial

INTRODUCTION TO IDRISI

Image Classification. Introduction to Photogrammetry and Remote Sensing (SGHG 1473) Dr. Muhammad Zulkarnain Abdul Rahman

Exercise 2-1 Cartographic Modeling

Introduction to ERDAS IMAGINE. (adapted/modified from Jensen 2004)

APPLICATION OF SOFTMAX REGRESSION AND ITS VALIDATION FOR SPECTRAL-BASED LAND COVER MAPPING

Aerial photography: Principles. Visual interpretation of aerial imagery

ENVI Tutorial: Basic Hyperspectral Analysis

Exercise 6-1 Image Georegistration Using RESAMPLE

SETTLEMENT OF A CIRCULAR FOOTING ON SAND

Introduction to SAGA GIS

AN INTEGRATED APPROACH TO AGRICULTURAL CROP CLASSIFICATION USING SPOT5 HRV IMAGES

Tutorial 7. Water Table and Bedrock Surface

ENVI Classic Tutorial: Basic Hyperspectral Analysis

Math 227 EXCEL / MEGASTAT Guide

Mapping 2001 Census Data Using ArcView 3.3

An Introduction To. Version Program Concept and Introduction Notes by David Landgrebe and Larry Biehl. MultiSpec Programming by Larry Biehl

Lab on MODIS Cloud spectral properties, Cloud Mask, NDVI and Fire Detection

Object Based Image Analysis: Introduction to ecognition

Remote Sensing & Photogrammetry W4. Beata Hejmanowska Building C4, room 212, phone:

Pre-Lab Excel Problem

General Digital Image Utilities in ERDAS

Glacier Mapping and Monitoring

v Working with Rasters SMS 12.1 Tutorial Requirements Raster Module Map Module Mesh Module Time minutes Prerequisites Overview Tutorial

Lab 3: Digitizing in ArcGIS Pro

F E A T U R E. Tutorial. Feature Mapping M A P P I N G. Feature Mapping. with. TNTmips. page 1

Chapter 17 Creating a New Suit from Old Cloth: Manipulating Vector Mode Cartographic Data

Spreadsheet Concepts: Creating Charts in Microsoft Excel

Introduction to Remote Sensing

Crop Counting and Metrics Tutorial

AEMLog Users Guide. Version 1.01

Intro to GIS (requirements: basic Windows computer skills and a flash drive)

COMBINING HIGH SPATIAL RESOLUTION OPTICAL AND LIDAR DATA FOR OBJECT-BASED IMAGE CLASSIFICATION

Remote Sensing & GIS (Bio/Env384 A): 10 November 2015 GIS Database Query

Homework 1 Excel Basics

GGR 375 QGIS Tutorial

Creating a Basic Chart in Excel 2007

INTRODUCTION TO GIS WORKSHOP EXERCISE

MODULE 1 BASIC LIDAR TECHNIQUES

ArcView QuickStart Guide. Contents. The ArcView Screen. Elements of an ArcView Project. Creating an ArcView Project. Adding Themes to Views

Classification. Vladimir Curic. Centre for Image Analysis Swedish University of Agricultural Sciences Uppsala University

Spectral Classification

COSC160: Detection and Classification. Jeremy Bolton, PhD Assistant Teaching Professor

Getting To Know The Multiform Bivariate Matrix

Roberto Cardoso Ilacqua. QGis Handbook for Supervised Classification of Areas. Santo André

v SMS Tutorials Working with Rasters Prerequisites Requirements Time Objectives

Geostatistics 3D GMS 7.0 TUTORIALS. 1 Introduction. 1.1 Contents

Tutorial 01 Quick Start Tutorial

Randy H. Shih. Jack Zecher PUBLICATIONS

The Gain setting for Landsat 7 (High or Low Gain) depends on: Sensor Calibration - Application. the surface cover types of the earth and the sun angle

Probabilistic Graphical Models Part III: Example Applications

RESEARCH ON THE VISUALIZATION SYSTEM AND APPLICATIONS OF UNCERTAINTY IN REMOTE SENSING DATA CLASSIFICATION

IDL Tutorial. Contours and Surfaces. Copyright 2008 ITT Visual Information Solutions All Rights Reserved

ENVI Tutorial: Map Composition

ENGRG Introduction to GIS

Error Analysis, Statistics and Graphing

Your Name: Section: INTRODUCTION TO STATISTICAL REASONING Computer Lab #4 Scatterplots and Regression

Introduction to the Google Earth Engine Workshop

Obtaining Submerged Aquatic Vegetation Coverage from Satellite Imagery and Confusion Matrix Analysis

Ex. 4: Locational Editing of The BARC

Microsoft Excel Using Excel in the Science Classroom

EXERCISE 2: GETTING STARTED WITH FUSION

Importing and processing a DGGE gel image

Learn the various 3D interpolation methods available in GMS

ENVI Classic Tutorial: Multispectral Analysis of MASTER HDF Data 2

Tutorial 3 Sets, Planes and Queries

Files Used in this Tutorial

Chapter 2 Surfer Tutorial

Lab 1: Exploring ArcMap and ArcCatalog In this lab, you will explore the ArcGIS applications ArcCatalog and ArcMap. You will learn how to use

ENVI Classic Tutorial: 3D SurfaceView and Fly- Through

IDL Tutorial. Working with Images. Copyright 2008 ITT Visual Information Solutions All Rights Reserved

Digitizing and Editing Polygons in the STS Gypsy Moth Project. M. Dodd 2/10/04

Introduction to Minitab 1

INCREASING CLASSIFICATION QUALITY BY USING FUZZY LOGIC

Application of nonparametric Bayesian classifier to remote sensing data. Institute of Parallel Processing, Bulgarian Academy of Sciences

Transcription:

FOR 324 Natural Resources Information Systems Lab #4 Introduction to Image Processing II and Map Accuracy Assessment (Adapted from the Idrisi Tutorial, Introduction Image Processing Exercises, Exercise 4-3 Supervised Classification) Name: Objectives: To learn how to digitize land use/land cover features from satellite imagery while the imagery is displayed onscreen To explore computer assisted land cover classification procedures, including both supervised and unsupervised techniques To learn how to conduct a map accuracy assessment Materials required: Idrisi GIS software program Seven bands of Landsat Thematic Mapper (TM) satellite imagery of ITHACA, NY What to submit for this lab: Answers to all questions, printouts of all classified images, and accuracy assessment error matrices for each classified image. Introduction In the last lab exercise, we drew the spectral response patterns for three kinds of land covers: urban, forest and water. We saw that the spectral response patterns of each of these cover types were unique. Land covers, then, may be identified and differentiated from each other by their unique spectral response patterns. This is the logic behind image classification. Many kinds of maps, including land-cover, soils, and bathymetric maps, may be developed from the classification of remotely sensed imagery. There are two methods of image classification: supervised and unsupervised. With supervised classification, the user develops the spectral signatures of known categories, such as urban and forest, and then the software assigns each pixel in the image to the cover type to which its signature is most similar. With unsupervised classification, the software groups pixels into categories of like signatures, and then the user identifies what cover types those categories represent. The steps for supervised classification may be summarized as follows: 1. Locate representative examples of each cover type that can be identified in the image (called training sites). 2. Digitize polygons around each training site, assigning a unique identifier to each cover type. 3. Analyze the pixels within the training sites and create spectral signatures for each of the cover types. 4. Classify the entire image by considering each pixel, one by one, comparing its particular signature with each of the known signatures. So-called hard classifications result from assigning each pixel to the cover type that has the most similar signature. Soft classifications, lab4.doc Page 1 of 13

on the other hand, evaluate the degree of membership of the pixel in all classes under consideration, including unknown and unspecified classes. Decisions about how similar signatures are to each other are made through statistical analyses. There are several different statistical techniques that may be used. These are often called classifiers. This exercise illustrates the hard supervised classification techniques available in IDRISI. 5. Assessing the accuracy of the classification using an error matrix. Procedures 1. Read through the complete lab assignment (this document) before you sit down in front of a computer. 2. Training Site Development We will begin by creating the training sites. The imagery we will classify is the area around Ithaca, NY, immediately surrounding the city. The satellite imagery we will use in this lab are the same imagery we saw in LAB #03, plus additional spectral bands of the same region (Figure 1). The training sites that will be created based on the knowledge of land cover types identified during a field visit. Ideally, we would send you out to the field to identify the land cover, but to save time and costs, we will interpret land cover classes from higher resolution digitial orthophoto quarter quad imagery (Figures 2(a)-(c)), whose outlines are highlighted on Figure 1. a) You will look at five known land-cover types, each will be assigned a unique integer identifier, and one or more training sites will be identified for each. The list of land cover types is given below, along with a unique identifier that will signify each cover type. While the training sites can be digitized in any order, they may not skip any number in the series, so if you have five different land-cover classes, for example, your identifiers must be 1 to 5. The suggested order (to create a logical legend category order) is: 1 Conifer forest 2 Hardwood forest 3 Agricultural land 4 Urban 5 Water b) Display the natural color, composite image created in the last lab (i.e., COMPOSITE123) (If you didn t save it, recreate the composite image using Display Composite, specifying ITHACA_TM1 as the blue image band, ITHACA_TM2 as the green image band and ITHACA_TM3 as the red image band.). Zoom in the region defined as AREA 1 on Figure 1 using the Zoom Window icon on the Tool Bar. c) Use the on-screen digitizing feature of IDRISI to digitize polygons around your training sites. On-screen digitizing in IDRISI is made available through the following three toolbar icons: Digitize Delete Feature Save Digitized Data lab4.doc Page 2 of 13

d) Select the Digitize icon from the toolbar. The following dialog box should open up. e) Enter TRAININGSITES as the name of the layer to be created. Use the Default Qualitative palette and choose to create polygons. Enter the feature identifier you chose for Conifer forest (e.g., 1). Click OK. The vector polygon layer TRAININGSITES is automatically added to the composition and is listed on COMPOSER. f) Your cursor will now appear as the digitize icon when in the image. Move the cursor to a starting point for the boundary of your training site and press the left mouse button. Then move the cursor to the next point along the boundary and press the left mouse button again (you will see the boundary line start to form). The training site polygon should enclose a homogeneous area of the cover type, so avoid including the roads and agricultural fields surrounding the conifer forest. Continue digitizing until just before you have finished the boundary, and then press the right mouse button. This will finish the digitizing for that training site and ensure that the boundary closes perfectly. The finished polygon is displayed with the symbol that matches its identifier. g) You can save your digitized work at any time by pressing the Save Digitized Data icon on the toolbar. Answer YES when asked if you wish to save changes. h) If you make a mistake and wish to delete a polygon, select the Delete Feature icon ( ). Select the polygon you wish to delete, then press the delete key on the keyboard. Answer YES when asked whether to delete the feature. You may delete features either before or after they have been saved. i) Use the navigation tools to zoom back out, then focus in on your next training site area, referring to Figures 1 & 2. Select the Digitize icon again. Indicate that you wish to Add features to the currently active vector layer. Enter an identifier for the new training site. Keep the same identifier if you want to digitize another polygon around the same cover type (e.g., if you want to digitize another CONIFER stand, enter 1 for the ID). Otherwise, enter a new identifier. lab4.doc Page 3 of 13

j) Any number of training sites, or polygons with the same ID, may be created for each cover type. In total, however, there should be an adequate sample of pixels for each cover type for statistical characterization. A general rule of thumb is that the number of pixels in each training set (i.e., all the training sites for a single land cover class) should not be less than ten times the number of bands. Thus, in this exercise, where we will use seven bands in classification, we should aim to have no less than 70 pixels per training set. k) Continue until you have training sites digitized for each of the five different land cover classes. Then save the file using the Save Digitized Data icon from the toolbar. 3. Signature Development a) After you have a training site vector file you are ready for the third step in the process, which is to create the signature files. Signature files contain statistical information about the reflectance values of the pixels within the training sites for each class. b) Run MAKESIG from the Image Processing/Signature Development (Image Processing Signature Development MAKESIG) menu. Choose Vector as the training site file type and enter TRAININGSITES as the file defining the training sites. Click the Enter Signature Filenames button. A separate signature file will be created for each identifier in the training site vector file. Enter a signature filename for each identifier shown (e.g., if your CONIFER training sites were assigned ID 1, then you might enter CONIFER as the signature filename for ID 1). When you have entered all the filenames, click on OK. Indicate that seven bands of imagery will be processed by pressing the up arrow on the spin button until the number 7 is shown. This will cause seven input name boxes to appear in the grid. Click the pick list button next to the first box and choose ITHACA_TM1 (the blue band). Click OK, then click the mouse into the second input box. The pick button will now appear on that box. Select it and choose ITHACA_TM2 (the green band). Enter the names of the other bands in the same way: ITHACA_TM3 (red band), and ITHACA_TM4 (near infrared band), ITHACA_TM5 (middle infrared band), ITHACA_TM6 (thermal infrared band) and ITHACA_TM7 (middle infrared band). Click OK when done. c) When MAKESIG has finished, open File IDRISI File Explorer from the menu. Click on the Signature file type (sig + sgf) and check that a signature exists for each of the five landcover types. If you forgot any, repeat the process described above to create a new training site vector file (for the forgotten cover type only) and run MAKESIG again. Exit from the IDRISI File Explorer. To facilitate the use of several subsequent modules with this set of signatures, we may wish to create a signature group file. Using group files (instead of specifying each signature individually) quickens the process of filling in the input information into module dialog boxes. Similar to a raster image group file, a signature group file is an ASCII file that may be created or modified with the Collection Editor. MAKESIG automatically creates a signature group file that contains all our signature filenames. This file has the same name as the training site file, TRAININGSITES. lab4.doc Page 4 of 13

d) Open File Collection Editor. From the Editor's menu, choose File Open. Then from the drop-down list of file types, choose Signature Group Files, and choose TRAININGSITES. Click Open. Verify that all the spectral reflectance signatures are listed in the group file and then close the Collection Editor. To compare the spectral reflectance signatures for these five landcover classes, we can graph them, just as we did by hand in the previous exercise. e) Run SIGCOMP (Image Processing Signature Development SIGCOMP) from the Image Processing/Signature Development menu. Choose to use a signature group file and choose TRAININGSITES. Display their means. Q. Part 3(e) Of the seven bands of imagery, which bands show the largest differences between the vegetative covers the best? f) Close the SIGCOMP graph, then run SIGCOMP again. This time choose to view only 2 signatures and enter the urban and the conifer signature files. Indicate that you want to view their minimum, maximum, and mean values. Notice that the reflectance values of these signatures overlap to varying degrees across the bands. This is a source of spectral confusion between cover types. Q. Part 3(f) Which of the two signatures has the most variation in reflectance values (widest range of values) in all of the bands? Why do you think this is? g) Another way to evaluate signatures is by overlaying them on a two-band scatterplot or scattergram. The scattergram plots the positions of all pixels on two bands, where reflectance of one band forms the X axis and reflectance of the other band forms the Y axis. The frequency of pixels at each X,Y position is signified by a quantitative palette color. Signature characteristics can be overlayed on the scattergram to give the analyst a sense of how well they are distinguishing between the cover types in the two bands that are plotted. To create such a display in IDRISI, use the module SCATTER. It uses two image bands as X and Y axes to graph relative pixel positions according to their values in these two bands. In addition, it creates a vector file of the rectangular boundary around the signature mean in each band that is equal to two standard deviations from this mean. Typically one would create and examine several scatter diagrams using different pairs of bands. Here we will create one scatter diagram using the red and near infrared bands. lab4.doc Page 5 of 13

h) Run SCATTER from the Image Processing/Signature Development menu (Image Processing Signature Development SCATTER). Indicate ITHACA_TM4 (the near infrared band) as the Y axis and ITHACA_TM5 (the middle infrared band) as the X axis. Give the output the name SCATTER and retain the default logarithm count. Choose to create a signature plot file and enter the name of the signature group file TRAININGSITES. Press OK. i) A note will appear indicating that a vector file has also been created. Click OK to clear the message, then bring the scatterplot image into focus.1 Choose Add Layer from Composer and add the vector file SCATTER with the Qual symbol file. Move the cursor around in the scatter plot. Note that the X and Y coordinates shown in the status bar are the X and Y coordinates in the scatterplot. The X and Y axes for the plot are always set to the range 0-255. Since the range of values in ITHACA_TM4 is 14-117 and that for ITHACA_TM5 is 9-255, most of the pixels are plotting in the lower left quadrant of the scatterplot. Zoom in on the lower left corner to see the plot and signature boundaries better. You may also wish to click on the Maximize Display of Layer Frame icon on the toolbar (or press the End key) to enlarge the display. j) The values in the SCATTER image represent densities (log of frequency) of pixels, i.e., the higher palette colors indicate many pixels with the same combination of reflectance values on the two bands and the lower palette colors indicate few pixels with the same reflectance combination. Overlapping signature boxes show areas where different signatures have similar values. SCATTER is useful for evaluating the quality of one's signatures. Some signatures overlap because of the inadequate nature of the definition of landcover classes. Overlap can also indicate mistakes in the definition of the training sites. Finally, overlap is also likely to occur because certain objects truly share common reflectance patterns in some bands (e.g., hardwoods and forested wetlands). It is not uncommon to go through several iterations of training site adjustment, signature development, and signature evaluation before achieving satisfactory signatures. For this exercise, we will assume our signatures are adequate and will continue on with the classification. 4. Classification Now that we have satisfactory signature files for all of our landcover classes, we are ready for the last step in the classification process to classify the images based on these signature files. Each pixel in the study area has a value in each of the seven bands of imagery (ITHACA_TM1-7). As mentioned above, these are respectively: Blue, Green, Red, Near Infrared, Middle Infrared, Thermal Infrared and another Middle Infrared bands. These values form a unique signature which can be compared to each of the signature files we just created. The pixel is then assigned to the cover type that has the most similar signature. There are several different statistical techniques that can be used to evaluate how similar signatures are to each other. These statistical techniques are called classifiers. We will create classified images with four of the hard classifiers that are available in IDRISI. The concepts underlining the various classification procedures will be illustrated using two-dimensional graphs (as if the spectral signature were made from only two bands). In the heuristic diagrams given on the following pages, the signature reflectance values are lab4.doc Page 6 of 13

indicated with lower case letters, the pixels that are being compared to the signatures are indicated with numbers, and the spectral means are indicated with dots. First, we will look at the parallelepiped classifier. This classifier creates boxes (i.e., parallelepipeds) using minimum and maximum reflectance values or standard deviation units (Z-scores) within the training sites. If a given pixel falls within a signature s box, it is assigned to that category. This is the simplest and fastest of classifiers and the option using Min/ Max values was used as a quicklook classifier years ago when computer speed was quite slow. It is prone, however, to incorrect classifications. Due to the correlation of information in the spectral bands, pixels tend to cluster into cigar- or zeppelin-shaped clouds. Figure A As illustrated in Figure A, the boxes become too encompassing and capture pixels that probably should be assigned to other categories. In this case, pixel 1 will be classified as deciduous (d s) while it should probably be classified as corn (c s). Also, the boxes often overlap. Pixels with values that fall at this overlap are assigned to the last signature, according to the order in which they were entered. a) All of the classifiers we will explore in this exercise may be found in the Image Processing Hard Classifiers menu. Run PIPED and choose the Min/Max option. Click the Insert Signature Group button and choose TRAININGSITES. Call the output image PIPED, and give a descriptive title. Press the Next button and retain all bands. Press OK. Note the zero-value pixels in the output image. These pixels did not fit within the Min/Max range of any training set and were thus assigned a category of zero. Examine the resulting land cover image. (Change the palette to Qual if necessary.) The next classifier we will use is a minimum distance to means classifier. This classifier calculates the distance of a pixel's reflectance values to the spectral mean of each signature file, and then assigns the pixel to the category with the closest mean. There are two choices on how to calculate distance with this classifier. The first calculates the Euclidean, or raw, distance from the pixel's reflectance values to each category's spectral mean. This concept is illustrated in two dimensions (as if the spectral signature were made from only two bands) in Figure B. Figure B Pixel 1 is closest to the corn (c's) signature's mean, and is therefore assigned to the corn category. The drawback for this classifier is illustrated by pixel 2, which is closest to the mean for sand (s s) even though it appears to fall within the range of reflectances more likely to be urban (u s). In other words, the raw minimum distance to mean does not take into account the spread of reflectance values about the mean. lab4.doc Page 7 of 13

b) Run MINDIST (the minimum distance to means classifier) and indicate that you will use the raw distances and an infinite maximum search distance. Click on the Insert Signature Group button and choose the TRAININGSITES signature group file. The signature names will appear in the corresponding input boxes in the order specified in the group file. Call the output file MINDISTRAW, enter a descriptive title, and click on the Next button. By default, all of the bands that were used to create the signatures are selected to be included in the classification. Click OK to start the classification. Examine the resulting land cover image. (Change the palette to Qual if necessary.) We will try the minimum distance to means classifier again, but this time with the second kind of distance calculation normalized distances. In this case, the classifier will evaluate the standard deviations of reflectance values about the mean creating contours of standard deviations. It then assigns a given pixel to the closest category in terms of standard deviations. We can see in Figure B that pixel 2 would be correctly assigned to the urban category because it is two standard deviations from the urban mean, while it is at least three standard deviations from the mean for sand. Figure B c) To illustrate this method, run MINDIST again. Fill out the dialog box in the same way as before, except choose the normalized option, and call the result MINDISTNORMAL. Enter a descriptive title for the image being created. Examine the resulting land cover image. (Change the palette to Qual if necessary.) The next classifier we will use is the maximum likelihood classifier. Here the distribution of reflectance values in a training site is described by a probability density function developed on the basis of Bayesian statistics (Figure C). This classifier evaluates the probability that a given pixel will belong to a category and classifies the pixel to the category with the highest probability of membership. Figure C d) Run MAXLIKE. Choose to use equal prior probabilities for each signature. Click the Insert Signature Group button, then choose the signature group file TRAININGSITES. The input grid will then automatically fill. Choose to classify all pixels and call the output image MAXLIKE. Enter a descriptive title for the output image. Press Next and choose to retain all bands and press OK. Maximum likelihood is the slowest of the techniques, but if the training sites are good, it tends to be the most accurate. lab4.doc Page 8 of 13

The final supervised classification module we will explore in this exercise is FISHER. This classifier is based on linear discriminate analysis. e) Run FISHER. Insert the signature group file TRAININGSITES. Call the output image FISHER and give a title. Press OK. f) Compare each of the classifications you created: PIPED, MINDISTRAW, MINDISTNORMAL, MAXLIKE, and FISHER. To do this, display all of them with the Default Qualitative palette. You may need to make the window frames smaller to fit all of them on the screen. Q. Part 4(f) Visually, which classification do you feel is 'best'? What specifically made you come to that conclusion? 5. Map Accuracy Assessment All classification procedures have produced different classified images. Which one is the most correct?? One way to assess the accuracy of the classification output is to directly compare the classified pixels to known, ground-truthed locations. In Step 2 of this lab, you created a TRAININGSITES file of known ground locations. We will use this to test the accuracy of the classification procedures. The approach used is to count the number of pixels correctly classified as water, urban, conifer, hardwood and agriculture that agree with our groundtruthing data, and how many were misclassified. Summarizing this counting is best presented in an Error Matrix. If an image is 100% correctly classified, than all cells in the error matrix will have a value of zero, except along the dialog. The larger the values in the cells off the dialog, the more the classified image has misclassified pixels. An error matrix includes column and row marginal totals, errors of omission and commission, an overall error measure, confidence intervals for that figure, and a Kappa Index of Agreement (KIA), both for all classes and on a per category basis. Errors of omission computes the proportion of cells of known covertype that were misclassified, while errors of commission computes the proportion of cells classified as a covertype that were not that particular covertype. The KIA is an index used to assess if the differences between two images are due to chance or real disagreement, and the value for the index can range from 0 (agreement from chance) to 1 (perfect true agreement). The coefficient is calculated using lab4.doc Page 9 of 13

observed accuracy chance agreement KIA = 1 chance agreement N xii = 2 N xi x x where x ii = number of combinations along the diagonal x i. = total observations in row i x. i = total observations in column i N = total number of cells i x i i a) Idrisi has a procedure for computing an Error Matrix, found under Image Processing Accuracy Assessment ERRMAT. Run ERRMAT. Insert the TRAININGSITES file as the Ground Truth Image and the PIPED file as the Categorical map image. Repeat for each of the other four classified images. Complete the table below. Classification procedure Overall KIA Landcover with Largest error Errors of omission Landcover with Smallest error Landcover with Largest error Errors of commission Landcover with Smallest error Parallelepiped Min. Distance (Raw) Min. Distance (Normalized) Max. Likelihood Fisher Q. Part 5(a) As a general conclusion, which one of the five covertypes is the most easily identified and accurately classified, and which covertype is the most misclassified? Q. Part 5(a) Which of the classification procedure(s) produced the best classified image? Which procedure(s) were not very successful in classifying the scene? As a final note, consider the following. If your training sites are very good, the Maximum Likelihood or FISHER classifiers should produce the best results. However, when training sites are not well defined, the Minimum Distance classifier with the standardized distances option often performs much better. The Parallelepiped classifier with the standard deviation option also performs rather well and is the fastest of the considered classifiers. lab4.doc Page 10 of 13