Background and Goal
In this lab, we learned how to evaluate the accuracy of classification
maps. This knowledge is important because it is a necessary post-processing exercise
that highlights the strengths and weaknesses of a classification map. The
results of accuracy assessments provide an easily read quality assessment and
rating of the map as a whole, as well as each class within the map.
Methods
Accuracy assessments were completed for both the
unsupervised classification model and the supervised classification model that
were discussed in Labs 4 and 5 respectively. The process is identical for both
assessments. The first step in this process is to generate random ground reference
testing samples. These samples are either collected through field work with a
GPS unit, or with a high resolution image. In this lab, we used a high
resolution image as our reference for these testing samples. The testing
samples were generated using ERDAS’ Accuracy Assessment tool. The
classification map was to be assessed, using the high resolution image as a
reference (this means that points would appear on the reference map, with the
classification being applied to the classification map). Normally the total
amount of testing samples should be equal to 50 for each class (meaning a
minimum of 250 for 5 classes), but for the sake of time we only used 125
samples for our 5 classes. We used a stratified random distribution parameter
to ensure a quality sample. Going further, we applied the use of minimum
points, making sure that each class had 15 samples at the very least (this
makes sure classes of smaller area are still factored into the accuracy
assessment). Figure 1 displays the random testing samples on the high
resolution image. Once the random testing samples were generated, each sample
was located and a classification was applied. This classification would then be
compared to the classification applied in the classification map in the
accuracy assessment, creating a matrix to display an accuracy assessment
report.
Figure 1: All of the test samples that were used in one of the Accuracy Assessments. The points were randomly generated to reduce bias, and there are 125 points total. |
Results
Figure 2 displays the accuracy assessment report of the unsupervised
classification map, while Figure 3 displays the accuracy assessment report of
the supervised classification map. The matrix displays the overall accuracy, Kappa
statistic, producer’s accuracy, and user’s accuracy. The rows in the matrix
list the classes that were actually classified on the classification map, while
the columns display the classes that were selected with the reference points. The
matrix thus displays producer’s accuracy (also known as omission error) and
user’s accuracy (also known as commission error). Overall accuracy is the
report of the overall proportion of correctly classified pixels in the image
based on the reference sample. The Kappa statistic is a measure of the
difference between observed agreement between two maps and the agreement that
might be attained by chance (it calculates to what degree chance has a role in
an accurate classification). A value below 0.4 means there is poor agreement
while a value above 0.8 means there is a strong agreement. The producer’s
accuracy is the accuracy for a given class that examines how many pixels on the
map are classified correctly. The user’s accuracy is the accuracy for a given
class that examines how many of the pixels on the classified image are actually
what they say they are. Overall, the unsupervised classification map is more
accurate, though neither maps are accurate enough to actually be used with
accuracy.
Sources
Earth Resources Observation and Science Center. (2000). [Landsat image used to create a classification map]. United States Geological Survey. Provided by Cyril Wilson.
National Agriculture Imagery Program. (2005). [High resolution image used for reference in accuracy assessment]. United States Department of Agriculture. Provided by Cyril Wilson.
No comments:
Post a Comment