Accuracy Assessment of Thematic Maps

Slides:



Advertisements
Similar presentations
Accuracy Assessment of Thematic Maps
Advertisements

AA_Vis_1. Students examine both the “true color” And “false-color Infrared” images provided by GLOBE AA_Vis_2.
Multiple Criteria for Evaluating Land Cover Classification Algorithms Summary of a paper by R.S. DeFries and Jonathan Cheung-Wai Chan April, 2000 Remote.
Geog 458: Map Sources and Errors Uncertainty January 23, 2006.
Accuracy Assessment Chapter 14. Significance Accuracy of information is surprisingly difficult to address We may define accuracy, in a working sense,
Lecture 14: Classification Thursday 18 February 2010 Reading: Ch – 7.19 Last lecture: Spectral Mixture Analysis.
AA_Vis_1. Using the software “MultiSpec,” students follow protocols to prepare a “clustered” image of their GLOBE Study Site.
Lecture 14: Classification Thursday 19 February Reading: “Estimating Sub-pixel Surface Roughness Using Remotely Sensed Stereoscopic Data” pdf preprint.
Spatial data quality February 10, 2006 Geog 458: Map Sources and Errors.
February 15, 2006 Geog 458: Map Sources and Errors
Copyright, © Qiming Zhou GEOG1150. Cartography Quality Control and Error Assessment.
Global Land Cover: Approaches to Validation Alan Strahler GLC2000 Meeting JRC Ispra 3/02.
Accuracy Assessment. 2 Because it is not practical to test every pixel in the classification image, a representative sample of reference points in the.
Chapter 9 Accuracy assessment in remotely sensed categorical information 遥感类别信息精度评估 Jingxiong ZHANG 张景雄 Chapter 9 Accuracy assessment in remotely sensed.
Methods of Validating Maps of Deforestation and Selective Logging Carlos Souza Jr. Instituto do Homem e Meio Ambiente da Amazônia—Imazon.
Diagnostic Metrics, Part 1 Week 2 Video 2. Different Methods, Different Measures  Today we’ll focus on metrics for classifiers  Later this week we’ll.
1 Evaluating Model Performance Lantz Ch 10 Wk 5, Part 2 Right – Graphing is often used to evaluate results from different variations of an algorithm. Depending.
Classification & Vegetation Indices
Using spectral data to discriminate land cover types.
Image Classification Digital Image Processing Techniques Image Restoration Image Enhancement Image Classification Image Classification.
Image Classification 영상분류
Accuracy Assessment Having produced a map with classification is only 50% of the work, we need to quantify how good the map is. This step is called the.
Computational Intelligence: Methods and Applications Lecture 12 Bayesian decisions: foundation of learning Włodzisław Duch Dept. of Informatics, UMK Google:
Accuracy of Land Cover Products Why is it important and what does it all mean Note: The figures and tables in this presentation were derived from work.
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Chuvieco and Huete (2009): Fundamentals of Satellite Remote Sensing, Taylor and Francis Emilio Chuvieco and Alfredo Huete Fundamentals of Satellite Remote.
Remote Sensing Classification Accuracy
Error & Uncertainty: II CE / ENVE 424/524. Handling Error Methods for measuring and visualizing error and uncertainty vary for nominal/ordinal and interval/ratio.
Chapter 8 Evaluating Search Engine. Evaluation n Evaluation is key to building effective and efficient search engines  Measurement usually carried out.
BOT / GEOG / GEOL 4111 / Field data collection Visiting and characterizing representative sites Used for classification (training data), information.
Environmental Modeling Validating GIS Models. 1. A Habitat Model Issues: ► Mapping Florida Scrub Jay habitat in the Kennedy Space Center in the Kennedy.
Reference data & Accuracy Assessment Dr. Russ Congalton Kamini Yadav.
Housekeeping –5 sets of aerial photo stereo pairs on reserve at SLC under FOR 420/520 –June 1993 photography.
Accuracy Assessment Accuracy Assessment Error Matrix Sampling Method
Accuracy Assessment of Thematic Maps THEMATIC ACCURACY.
Chapter 10 Confidence Intervals for Proportions © 2010 Pearson Education 1.
26. Classification Accuracy Assessment
9.3 Hypothesis Tests for Population Proportions
Confidence Intervals for Proportions
Confidence Intervals for Proportions
Land Cover Mapping and Habitat Analysis
26. Classification Accuracy Assessment
Chapter 21 More About Tests.
HIERARCHICAL CLASSIFICATION OF DIFFERENT CROPS USING
Classification of Remotely Sensed Data
ASTER image – one of the fastest changing places in the U.S. Where??
Gauss-Siedel Method.
Confidence Intervals for Proportions
Confidence Intervals for Proportions
Other Kinds of Arrays Chapter 11
Map of the Great Divide Basin, Wyoming, created using a neural network and used to find likely fossil beds See:
Chapter 25 Comparing Counts.
Land Cover Mapping and Habitat Analysis
Incorporating Ancillary Data for Classification
More about Tests and Intervals
Autar Kaw Benjamin Rigsby
The GISCO task force “Remote Sensing for Statistics”
REMOTE SENSING Multispectral Image Classification
Housekeeping 5 sets of aerial photo stereo pairs on reserve at SLC under FOR 420/520 June 1993 photography.
Assessment of data quality
network of simple neuron-like computing elements
Uncertainty and Error
ALI assignment – see amended instructions
Chapter 26 Comparing Counts.
How to create a GREAT experiment
11E The Chi-Square Test of Independence
Confidence Intervals for Proportions
Chapter 26 Comparing Counts.
Confidence Intervals for Proportions
More on Maxent Env. Variable importance:
Presentation transcript:

Accuracy Assessment of Thematic Maps Thematic Accuracy

Learning Objectives What is accuracy assessment, and why is it so important? What is the typical procedure for assessing thematic accuracy of maps? What are reference data, and how are they collected? What is contingency table analysis, and how do you calculate the common accuracy metrics? As a map maker or user, what can you learn from accuracy assessment?

What is Accuracy Assessment? Map Accuracy -- The proportion of agreement between a classified map and reference data assumed to be correct. Thematic Precision – The level of detail that is mapped. A map distinguishing lodgepole pine, Douglas fir, Ponderosa, etc. is more precise (but probably less accurate!) than a map just showing Forest. Accuracy and precision are DIFFERENT!

Spatial Accuracy vs. Thematic Accuracy Thematic accuracy is how well the class names on the map correspond to what is really on the ground. Spatial accuracy quantifies errors in the locations of boundaries. Thematic and spatial accuracy are related but are usually treated separately.

Steps for performing thematic accuracy assessment: Develop a sampling scheme Collect reference data Compare reference data to classified map Compute accuracy metrics and deliver to map users

Sampling Schemes Reference locations must be unbiased Reference locations must be large enough to find with certainty on your classified image You must ensure that you can correctly identify the types at your reference sites You must visit LOTS of reference locations

Reference data collection sites for 2001 NLCD map, with ecoregions.

Collecting Reference Data Field collection Use GPS to find pre-determined reference sites Use explicit definitions of the classes to insure that reference data are consistent (should have done this with map classes too) Develop efficient plan for visiting sites and contingencies for unreachable sites

National Park Service image (nps.org)

Collecting Reference Data (cont.) Double Sampling – Using high resolution remotely sensed data to assess the accuracy of low resolution data Air photos, air videography, higher resolution satellites, etc. Requires some ground work to “train” yourself to accurately interpret the reference data May negate the assumption of accurate reference data! Relatively inexpensive, good for inaccessible areas, fast, etc.

Air Videography

Quantifying Accuracy – Comparing Mapped Types to Reference Types Contingency Tables = Error Matrices = Confusion Matrices Traditional Accuracy Statistics are calculated using the error matrix

Contingency Table or Error Matrix Error matrix is an n x n array where n is the number of classes Rows: reference (“correct”) data (n rows) Columns: mapped classes (n cols) Note that rows and columns can be switched, so you have to pay attention!

Error Analyses - Example Consider a map with 3 thematic classes (water, forest, and urban) Collect 95 reference sites (Water 33, Forest 39, and Urban 23) Compare reference locations to the same places in the map Create error matrix (contingency table).

Contingency Table Reference data Classified image Forest Urban Total Water Forest Urban Total 21 5 7 33 6 31 2 39 1 22 23 27 37 95

Classification Accuracy Lots of ways to look at the thematic accuracy of a classification Overall accuracy Errors of omission Errors of commission User’s accuracy Producer’s accuracy Accuracy statistics (e.g., Kappa) Fuzzy accuracy

Overall Accuracy Of all of the reference sites, what proportion were mapped correctly? Easiest to understand but least amount of information for map users and map producers (us).

Overall Accuracy Correctly classified: 21 + 31 + 22 = 74 Reference data Classified image Water Forest Urban Total 21 5 7 33 6 31 2 39 1 22 23 27 37 95 Correctly classified: 21 + 31 + 22 = 74 Total number reference sites = 95 Overall accuracy = 74 / 95 = 77.9%

Off-diagonal Elements The off-diagonal elements of a contingency table tell us the most about how to improve our remote sensing classification! Should spend lots of time examining ERRORS to figure out what went wrong

Errors of Omission The type on the ground is not that type on the classified image – the real type is OMITTED from the classified image.

Omission Error For water: 5 + 7 = 12 12 / 33 = 36% For forest: Reference data Classified image Water Forest Urban Total 21 5 7 33 6 31 2 39 1 22 23 27 37 95 For water: 5 + 7 = 12 12 / 33 = 36% For forest: 6 + 2 = 8 8 / 39 = 20% For urban: 0 + 1 = 1 1 / 23 = 4%

Errors of Commission A type on the classified image is not that type on the ground – the type is COMMITTED to the classified image.

Commission Error For water: 6 + 0 = 6 6 / 27 = 22% For forest: Reference data Classified image Water Forest Urban Total 21 5 7 33 6 31 2 39 1 22 23 27 37 95 For water: 6 + 0 = 6 6 / 27 = 22% For forest: 5 + 1 = 6 6 / 37 = 16% For urban: 7 + 2 = 9 9 / 31 = 29%

Producer’s Accuracy Map accuracy from the point of view of the map maker (PRODUCER). How often are real features on the ground correctly shown on the map?

Producer’s Accuracy Can be computed (and reported) for each thematic class Producer accuracy = 100 – omission error Water: 100 – 36 = 64% Forest: 100 – 20 = 80% Urban: 100 – 4 = 96% Or PA = #correct/row total

User’s Accuracy Accuracy from the point of view of a map USER (not a map maker). How often is the type the map says should be there really there?

User’s Accuracy Can be computed (and reported) for each thematic class User accuracy = 100 – commission error Water: 100 – 22 = 78% Forest: 100 – 16 = 84% Urban: 100 – 29 = 71% Or UA = #correct/column total

Kappa Statistic A measure of how accurate your map is, taking into account the accuracy expected by chance alone Affected by omission, commission, and overall accuracy Frequently reported in the literature and used to compare maps

Kappa Statistic Kappa = (observed – expected)/(1 – expected) Observed is the overall accuracy from your contingency table (the sum of the diagonal)/(Total reference pixels) Expected is calculated from a matrix of “marginal products” Create a matrix by multiplying marginal totals. Expected = diagonal sum/total sum (from the marginal table) (See Campbell for calculation details)

Fuzzy Accuracy Recognizes that binary accuracy assessment (right vs. wrong) discards important information Can use scale of “rightness” or “wrongness” to get more nuanced accuracy assessment (e.g. absolutely right, good, reasonable, mostly wrong, absolutely wrong) There are many fuzzy metrics that can capture this information numerically.

Accuracy Assessment -- Summary Critical for any mapping project Expensive – should be included in the budget from the start Requires collection of accurate reference data, either in the field or from higher resolution data Should include many measures of accuracy to give users more information about the product.