Upload
polylsgiedx
View
436
Download
0
Embed Size (px)
Citation preview
1
LSGI332 Remote sensing
Lecture 11. Image ClassificationLillesand and Keifer 6th edition pp.545-581, 585-592
2
Image classification• Objective: automatically categorise all pixels using
numerical, Spectral Pattern for each pixel NB. ‘pattern’ refers to Spectral Pattern, not spatial
• Spectral Pattern Recognition includes a range of classification procedures
• Usually refers to Land Cover Mapping• Uses more than one band:otherwise = Density Slice• Comprises 2 main approaches:
– Supervised, and – Unsupervised
3
Classified SPOT image of Hong Kong by Supervised classification method
Note scattered water pixels- what causes these?
4
Image classification: spectral pattern recognition
Fig. 7.18
5
Supervised classification
• Operator controlled• There are three distinct stages:
– Training– Classification– Output
6
Supervised classification
Fig. 7.17
7
Training stage in supervised classification• Training stage – analyst identifies representative training areas – develops
numerical description of spectral attributes of each cover type in scene• Analyst must develop training statistics for all spectral classes constituting
each information class to be discriminated by the classifier
Examples‘water’ may be represented by 4 or 5 spectral classes‘agriculture’ may contain several crop types and each crop type may contain several spectral classes due to different planting dates, moisture conditions etc.
8
Training area creationmain objective – to assemble a set of descriptive statistics
training areas must be representative and complete‘The point that must be emphasised is that all spectral classes constituting each information class must be adequately represented in the training set statistics used to classify an image’ (Lillesand and Keifer)
9
Training area statistics for one class, 6 bands
10
Sample histograms for
pixels in training areas for a cover
type ‘hay’
11
Coincident spectral plots for training
areas in 5 bands for 6 cover types
Very useful for examining the data before classifying
Thermal band often very useful if optical bands have overlap
12
Classification stage
• each pixel in dataset categorised into the land cover class it most closely resembles
• if not similar to any, labelled ‘unknown’
• category label set to ‘output image’
• every pixel is classified – a pixel-by-pixel operation (unlike image enhancement and image filtering
13
Classification methods
• Several different approaches:– Minimum-distance-to-means classifier– Parallelepiped classifier– Gaussian maximum likelihood classifier– Neural network
• Increasing sophistication
14
Band 2
Ban
d 1
0
2550
255
TM Band 3
TM B
and
4concretehigh buildingsgrass slopewaterbare soilsforest
Scatterplots for training areas from SPOT image of Hong Kong
Pixel in 2-Dimensional space
Training area pixels in 2-D space: should be visualised as n-dimensional
15
Scattergram of training areas• scattergram
illustrating the pixel observations of six cover types
0
2550
255
TM Band 3
TM B
and
4
concretehigh buildingsgrass slopewaterbare soilsforest
Cover Type Colour No. Points
Water Cyan 3793
Concrete Purple 975
High buildings Thistle 1866
Bare soils Coral 784
Grass slope Yellow 924
Forest Green 3122
Water
High buildings
Concrete
Bare soils
Forest
16
Scattergram of real image• Spectral ranges of the training areas on the scattergram (right)
and the image classes to which they belong (left).
TM Band 3
8
28
49
69
90
14 35 56 77 88
TM B
and
4
Concrete
Water
17
Scatter plot for training areas
Source: Lillesand and Keifer 6th ed.
18
Minimum Distance to Means Classifier
19
Scatterplot training area pixels for SPOT image of Hong Kong, with cluster means
0
2550
255
TM Band 3
TM B
and
4
concretehigh buildingsgrass slopewaterbare soilsforest
1 2
20
Parallelpiped Classifier
21
Scatterplot training area pixels for SPOT image of Hong Kong with boxes drawn for range
0
2550
255
TM Band 3
TM B
and
4
concretehigh buildingsgrass slopewaterbare soilsforest
1
2
22
Parallel Piped Classifier with stepped class region boundaries
23
Equiprobability Contours defined by Maximum Likelihood Classifier
24
Probability density function for Maximum Likelihood Classifier
25
Equiprobability contours plotted against scatterplot of Hong Kong training area data
0
2550
255
TM Band 3
TM B
and
4
concretehigh buildingsgrass slopewaterbare soilsforest
1 2
26
Comparison of SPOT image classified by Maximum Likelihood, and Minimum Distance to means classifiers
27
Unsupervised classification• Software controlled• Examine all bands for each pixel in an image - the
process is iterative• Aggregate them into a number of classes based on natural
groupings or clusters present in the image values• Result in spectral classes (clusters) and initially, their
identity are unknown• Each class is then identified from knowledge or reference
data• Operator intervention only after data has been placed into
classes
28
Menu for Unsupervised classification
29
ISODATA classification of part of Guangdong
30
Classes resulting from Unsupervised Classification
31
Classification and post processing• Classified image before (left) and after (right)
post-classification process (majority filter)
Water Concrete High buildings Bare soils Grass slope Forest
32
Problems of land use classification in urban areas
• Similar spectral response of urban surfaces• Bi-modal spectral response of urban land use
features• Land Cover more useful concept than Land Use
33
A.The nature of urban areas affecting image interpretation
Source: Remote Sensing of the environment by J.R. Jensen, 2000
(i) Spectral characteristics of common urban materials
34
Appearance of tree canopy: colour infra-red air photo and IKONOS multispectral images
35
Product validation: Accuracy assessment
• Error matrix (or confusion table)– Comparison of the relationship between known
reference data and the corresponding results of the classification
– It is common to average the correct classifications and regard this as the overall classification accuracy (in this case 81%)
– omission errors shown in column and commission errors shown in row
36
Example of confusion matrixTraining data (known cover types)
Classificationresults
Water Concrete Highbuildings
Baresoils
Grassslopes
Forest Rowtotal
Water 93 0 2 1 0 0 96
Concrete 0 65 4 6 0 0 75
High buildings 2 3 124 5 9 12 155
Bare soils 2 3 21 165 24 12 227
Grass slopes 0 0 6 16 201 45 268
Forest 0 0 8 9 76 512 605
Column total 97 71 165 202 310 581 1426
Producer’s accuracy User’s accuracy
W = 93/97 = 96% B = 165/202 = 82% W = 93/96 = 97% B = 165/227 = 73%
C = 65/71 = 92% G = 201/310 = 65% C = 65/75 = 87% G = 201/268 = 75%
H = 124/165 = 75% F = 512/581 = 88% H = 124/155 = 80% F = 512/605 = 85%
Overall accuracy = (93 + 65 + 124 + 165 + 201 + 512)/ 1426 = 81% = (1160 - 365.11) / (1426 - 365.11) = 0.749
• 124 sample points have been correctly classified as high buildings• But 2 genuine high building samples have been classified as water, • and 2 water samples have been classified as high buildings
37
Interpretation of confusion matrix
• Producer’s accuracy– Correct classified pixels / column total (total pixels of
pixels allocated to the class), e.g. 124/165=75%– Meaning: There are 165 pixels which are actually
High Buildings and only 124 of those are correctly assigned to High Buildings
• User’s accuracy– Correct classified pixels / row total (total number of
pixels classified as a class), e.g. 124/155=80%– Meaning: A total 155 pixels have been classified as
High Buildings but only 124 pixels are actually High Buildings
38
Accuracy assessment: k coefficient• Global assessment of classification accuracy:
Cohen’s kappa coefficient ()• The optimal score is 1.0 (perfect classification). In
previous example, N = 1426, d = 1160, q = 520609, and = 0.749.
n
jiji xx
1,
n
ijij xx
1,
i.e. the sum over all columns for row i
i.e. the sum over all rows for column j
qNqdN
2
*
N = total number of samples, d = total number of cases in diagonal cells of the error matrixq=x+,j*x,i+
39
Output stage
• production of thematic maps, tables or statistics
• the classification output becomes a GIS input
• needs knowledge of area and/or reference data