Digital Image Classification
Digital Image Classification
POONAM S. TIWARI
Photogrammetry and Remote Sensing
Division
Digital Image Classification Poonam S.Tiwari
iirs
Main lecture topics
• What is it and why use it?
• Image space versus feature space
• Distances in feature space
• Decision boundaries in feature space
• Unsupervised versus supervised training
• Classification algorithms
• Validation (how good is the result?)
• Problems
COMPUTER ASSISTED
• mainly spectral information
STRATIFIED
• using GIS functionality to incorporate
• knowledge from other sources of information
Mean Vector µI =
PIXEL A: 34,25
PIXEL B: 34,24
PIXEL C: 11,77 A
B
Input layer
Feature vectors
Low frequency
High
frequency
band y
. .
..
(units of 5 DN)
Max y
. ... .
(0,0) band x (units of 5 DN) Min y ..
Euclidian distance
(0,0) Min x Max x
Cluster
Where:
D = spectral distance
n = number of bands (dimensions)
i = a particular band
di = data file value of pixel d in band i
ei = data file value of pixel e in band i
This is the equation for Euclidean distance—in two dimensions (when n = 2), it
can be simplified to the Pythagorean Theorem (c2 = a2 + b2), or in this case:
D2 = (di - ei)2 + (dj - ej)2
Steps in supervised
classification
• Identification of sample
areas (training areas)
• Partitioning of the A class sample
feature space • Is a number of
training pixels
•Forms a cluster in
feature space
A cluster
• Is the representative for
a class
• Includes a minimum
number of observations
(30*n)
• Is distinct
Digital Image Classification Poonam S.Tiwari
iirs
UNSUPERVISED CLASSIFICATION
• The identities of land cover types to be specified
as classes within a scene are generally not
known a priori because ground reference
information is lacking or surface features within
the scene are not well defined.
• The computer is required to group pixels with
similar spectral characteristics into unique
clusters according to some statistically
determined criteria.
• Analyst then combine and relabels the spectral
clusters into information classes.
class a
• decide on
decision
class c boundaries
1. Euclidian Distance:
Where:
D = spectral distance
n = number of bands (dimensions)
i = a particular band
di = data file value of pixel d in band i
ei = data file value of pixel e in band i
2. Divergence
Band 2 Band 2
0 0
0 Band 1 255 0 Band 1 255
Class “unknown”
DISADVANTAGES:
Since parallelepiped has corners, pixels that are actually quite far,
spectrally from the mean of the signature may be classified
Parallelepiped Corners
Compared to the
Signature Ellipse
300
200
100
Decision rule:
Priority to the shortest distance to the class mean
"Unknown"
0
Band 2 0 Band 1 255
255
0
0 Band 1 255
Band 2
0
0 Band 1 255
Threshold Distance
Pixels 1 and 2 now represent cluster #1. Note that the location of cluster 1 has
migrated from 10,10 to 15,15 after the first iteration. Now, pixel 3 distance
(D=15.81) is computed to see if it is greater than the minimum threshold, R=15. It
is, so pixel location 3 becomes cluster #2. This process continues until all 20
clusters are identified. Then the 20 clusters are evaluated using a distance measure,
C (not shown), to merge the clusters that are closest to one another.
ISODATA does not allocate its initial mean vectors based on the
analysis of pixels rather, an initial arbitrary assignment of all Cmax
clusters takes place along an n-dimensional vector that runs between
very specific points in feature space.
Digital Image Classification Poonam S.Tiwari
iirs
ISODATA algorithm normally requires the analyst
to specify-
Cmax : maximum no. of clusters to be identified.
T:maximum % of pixels whose class values are
allowed to be unchanged between iterations.
M :maximum no. of times isodata is to classify pixels
and recalculate cluster mean vectors.
Minimum members in a cluster
Maximum standard deviation for a cluster.
Split separation value (if the valuse is changed from 0.0, it takes
the place of S.D. )
Minimum distance between cluster means.
a) Distribution of 20 ISODATA
mean vectors after just one
iteration
b) Distribution of 20 ISODATA
mean vectors after 20 iterations.
The bulk of the important
feature space (the gray
background) is partitioned rather
well after just 20 iterations.
These include:
– 3) The final point here, in terms of the potential for generation of error
is the mislabeling of classes. The most obvious example of this is to
label a training site water when in fact it is something else. This will
result in, at best a skewing of your class statistics if your training site
samples are sufficiently large, or at worst shifting the training
statistics entirely if your sites are relatively small.
• Of the three the systematic sample is the least useful. This approach to
sampling may result in a sample distribution which favours a particular class
depending on the distribution of the classes within the map.
• The truly random strategy however may not yield a sample design that covers
the entire map area, and so may be less than ideal.
• In many instances the stratified random sampling strategy is the most useful
tool to use. In this case the map area is stratified based on either a
systematic breakdown followed by a random sample design in each of the
systematic subareas, or alternatively through the application of a random
sample within each of the map classes. The use of this approach will ensure
that one has an adequate cover for the entire map as well as generating a
sufficient number of samples for each of the classes on the map
Digital Image Classification Poonam S.Tiwari
iirs Sample Size:
• which, for an error matrix set up with the row and column
assignments as stated, is computed as the user accuracy
For an error matrix with r rows, and hence the same number of columns, let
– A = the sum of r diagonal elements, which is the numerator in the
computation of overall accuracy
– Let B = sum of the r products (row total x column total).
• Then
Thus