Professional Documents
Culture Documents
ANALYSIS
TYPES OF CLASS
Information classes are those categories of interest
that the analyst is actually trying to identify in the
imagery, such as different kinds of crops, different forest
types or tree species, different geologic units or rock
types, etc.
Spectral classes are groups of pixels that are uniform (or
near-similar) with respect to their brightness values in
the different spectral channels of the data.
The objective is to match the spectral classes in the data
to the information classes of interest.
However, it is rare that there is a simple one-to-one
match between these two types of classes.
TYPES OF CLASS
Many times it is found that 2 to 3 spectral classes merge
to form one informational class, while some classes may
not be of any particular interest.
It is the analysts job to decide on the utility of the
different spectral classes and their correspondence to
useful information classes.
IMAGE CLASSIFICATION
AND ANALYSIS
Common classification procedures can
be broken down into two broad
subdivisions based on the method used:
i. supervised classification
ii. unsupervised classification.
SUPERVISED CLASSIFICATION
In a supervised classification, the analyst identifies in the
imagery, homogeneous representative samples of the
different surface cover types (information classes) of
interest.
These samples are referred to as training areas.
The selection of appropriate training areas is based on
the analysts familiarity with the geographical area and
knowledge of the actual surface cover types present in
the image.
Thus, the analyst is supervising the categorization of a
set of specific classes.
UNSUPERVISED CLASSIFICATION
In essence, it is reverse of the supervised classification process.
Spectral classes are grouped, first, based solely on the
numerical information in the data, and are then matched by the
analyst to information classes (if possible).
Programs called clustering algorithms are used to determine the
natural groupings or structures in the data.
Usually, the analyst specifies how many groups or clusters are
to be looked for in the data.
In addition to specifying the desired number of classes, the
analyst may also specify parameters related to the separation
distance amongst the clusters and the variation within each
cluster.
UNSUPERVISED CLASSIFICATION
The final result of this iterative clustering process may
result in some clusters that the analyst would like to
subsequently combine, or that some clusters have been
broken down, each of these require a further iteration of
the clustering algorithm.
Thus, unsupervised classification is not completely
without human intervention. However, it does not start
with a pre-determined set of classes as in a supervised
classification.
SUPERVISED CLASSIFICATION
In order to carry out supervised classification the analyst
may have to adopt a well defined procedure in so as to
achieve a satisfactory classification of information.
The important aspects of conducting a rigorous and
systematic supervised classification of remote sensor data
are as follows:
CLASSIFICATION SCHEME
Classification schemes have been developed so that
they can readily be incorporated as meanful land use
and land cover data as obtained by interpreting remote
sensing data.
Some of the important are
U.S. Geological Survey
Classification System,
Land
Use/Land
Cover
Level II
11 Residential
12 Commercial and services
13 Industrial
2. Agricultural
land
3. Rangeland
4. Forest land
52 Lakes
53 Reservoirs
54 Bays and estuaries
6. Wetland
61 Forested wetland
62 Non-forested wetland
71 Dry salt flats
72 Beaches
73 Sandy areas other than beaches
7.Barren land
8.Tundra
82 Herbaceous tundra
83 Bare ground
84 Mixed tundra
91 Perennial snowfields
identify and select sites within the image that are representative
of the land cover classes of interest.
Training data should be of value if the environment from which
they obtained is relatively homogenous.
The image coordinates of these sites are identified and used to
extract statistics from the multispectral data for each of these
areas.
For each feature class c, the mean value (ci) for each and
variance-covariance matrix (Vc) are calculated in a similar
manner as explained.
The success of a supervised classification depends upon the
training data used to identify different classes.
Hence selection of training data has to be done meticulously
keeping in mind each training data set has some specific
characteristics.
These characteristics are discussed below.
Band 1
Agriculture - 1
Barren Land
Band 2
Band 3
Band 4
Band 5
Band 7
Layer
Minimum
90.000
81.000
95.000
89.000
111.000
96.000
Maximum
122.000
121.000
142.000
111.000
149.000
140.000
Mean
104.985
101.685
121.108
97.870
128.649
120.696
Std. Dev
4.080
4.911
6.759
3.693
5.427
6.253
16.650
16.706
24.118
21.533
29.809
45.680
9.714
14.328
20.485
13.641
13.946
19.867
28.490
16.153
29.457
18.071
25.354
34.853
17.607
30.133
39.097
Feature Selection:
Once the training statistics have been systematically collected
from each band for each class of interest, a judgment must be
made to determine those bands that are most effective in
discriminating each class from all others.
This process is commonly called feature selection.
The goal is to delete from the analysis those bands that provide
only redundant spectral information.
In this way the dimensionality (i.e. the number of bands to be
processed) in the data set may be reduced.
This process minimizes the cost of the digital image classification
(but hopefully, not the accuracy).
Feature Selection
One dimensional decision
boundary
CLASS 1
No. of pixels
CLASS 2
n
n!
C
q q! n q !
Thus, if there are six thematic mapper bands and we
are interested in the three best bands to use in the
classification this results in 20 combinations that must
be evaluated.
6!
6
C
3 3! 6 3!
720
= 20 combinations
6(6)
But what about the case where there are more than two
classes? In this instance, the most common
solution is to
E
compute the average divergence, Divergavg.
This involves computing the average over all possible pairs of
classes, c and d, while holding the subset
of bands, q constant.
0.5Tr [(V V )( M M )( M M ) )]
Then another subset of bands, q is selected for the m classes
and analyzed.
1
c
1
d
m 1
m 1
c 1
c 1
Diverg
c
cd
Diverg
avg
( Diverg cd )
2000{1 exp
}
8
Feature Selection
The best q features are those dimensions whose sum of the
Bhattacharyya distance between the m(m-1)/2 classes is
highest.
The JM distance (also sometimes called
0.5Tr [(V Vthe
)( M Bhattacharrya
M )( M M ) )]
distance ) between a pair of probability distributions (spectral
classes) is defined as
1
c
J ij {
1
d
p ( x | i )}2 dx
SELECTION OF APPROPRIATE
CLASSIFICATION ALGORITHM
Various supervised classification methods have been used to
assign an unknown pixel to any one of the classes.
The choice of a particular classifier or decision rule depends on
the nature of the input data and the desired output.
Parametric classification algorithms assume that the observed
measurement vectors Xc obtained for each class in each
spectral band during the training phase of the supervised
classification are Gaussian in nature (i.e., they are normally
distributed).
Nonparametric classification algorithms make no such
assumption.
Among the most frequently used classification algorithms are
the Minimum Distance, Parallelepiped, and Maximum
Likelihood classifier.
Dist
BV
ijk
ck BVijl cl
2
The use of priori probability helps in incorporating the effects of relief and
other terrain characteristics. The disadvantage of this classifier is that it
requires a large computer memory space and computing time, and yet
sometimes may not produce the best results.
CLASSIFICATION ACCURACY
ASSESSMENT
No classification task using remote sensing data is
complete till an assessment of accuracy is performed.
The analyst and the user of a classified map would like to
know as to how accurately the classes on the ground
have been identified on the image.
The term accuracy correlates to correctness.
In digital image processing, accuracy is a measure of
agreement between standard information at given
location to the information at same location on the
classified image.
Generally, the accuracy assessment is based on the
comparison of two maps; one based on the analysis of
remote sensing data and second based on information
derived from actual ground also known as the reference
map.
Error Matrix
The standard form for reporting site-specific accuracy is
the error matrix, also known as the confusion matrix or
the contingency table.
An error matrix not only identifies the overall error for
each category, but also the misclassification for each
category.
An error matrix is essentially consists of an n by n array,
where n is the number of class or categories on the map
reference.
Here the rows of the matrix represent the true classes or
information on the reference map, while column of the
matrix represent the classes as identified on the classified
map.
Error Matrix
Class
Classified Image
Urban
Crop
Range
Water
Forest
Barren
Total
Crop
Row Marginal
Reference Image
Urban
Range
Water
Forest
Barren
Total
Column Marginal
Each row shows errors of omission while each column shows errors of
commission
Correctly identified pixels
Total sum of correctly identified pixels
Error Matrix
The values in the last column gives the total number
of true points per class used for assessing the
accuracy.
Similarly, the total at the bottom of each column
gives information regarding the number of
points/pixels per class in the classified map.
The diagonal elements of the error matrix indicate
the number of points/pixels correctly identified both
the reference and classified maps.
The sum total of all these diagonal elements is
entered at the right hand side bottom most element,
i.e., total number of points/ pixels correctly classified
both in the reference and classified maps.
ERROR MATRIX
The off-diagonal elements of the error matrix provide
information on errors of ommission and commission
respectively.
Errors of ommission are found in the upper right half of
the matrix and for each class it is computed by taking the
sum of all the non-diagonal elements along each row, and
dividing it by the row total of that class.
ACCURACY INDICES
Measure
Overall
accuracy
Users
accuracy
Producers
accuracy
Abbreviation
OA
Explanation
Percent of samples
correctly classified
1
N
Story and
Congalton
(1986)
nii N i
Story and
Congalton
(1986)
nii M i
Story and
Congalton
(1986)
i 1
UA
Index of individual
class accuracy
computed from row
total
PA
Index of individual
class accuracy
computed from
column total
AAu
1
q
nii
Ni
1
q
nii
Mi
Average
accuracy
AAp
Base
reference (s)
Formula
i 1
q
i 1
ii
Fung and
LeDrew
(1988)
ACCURACY INDICES
Weighted
Kappa
CAp
Average of overall
accuracy and average
users accuracy
1
OA AAp
2
Proportion of
agreement after
removing the proportion of agreement by
chance
Po Pe
1 Pe
Kw
Proportion of
weighted
disagreement
corrected for chance
Combined
accuracy
Kappa
coefficient
of agreement
CAu
Average of overall
1
OA AA p
accuracy and average
2
users accuracy
1
v
ij
Poij
ij
Peij
Fung and
LeDrew
(1988)
Congalton et
al. (1983)
Rosenfield &
FitzpatrickLins (1986)
ACCURACY INDICES
Conditional
Kappa
K+I
Conditional Kappa
computed from the ith Po ( i ) Pe(i 1)
column in error matrix
1 Pe( i )
(Producers)
Te
Tau
coefficient
Tp
Conditional
Tau
Ti+
T+I
1
q
1
1
q
Po
Po Pr
1 Pr
Foody
(1992),
Ma and
Raymond
(1995)
Po (i ) Pi
1 Pi
Po ( i ) Pi
1 Pi
Naesset
(1996)
Error Matrix
Actual
class
Predicted class
Heather
Water
Bare soil
Pasture
Total
826
27
858
Water
878
878
Forest 1
720
183
910
Forest 2
33
21
878
934
Bare soil
61
560
621
Pasture
219
220
920
878
741
1067
589
226
4081
Heather
Total
Forest 1 Forest 2
ERROR MATRIX
(p' )(q ) 50
p p' 1.645
n
n
where
p = Overall accuracy at 95% confidence level
p'= the overall accuracy,
q = 100- p', and
n = the sample size.
If this value of p exceeds the defined criterion at the lower limit,
then it is possible to accept this classification with 95%
confidence limit.
Normally, the defined criterion for confidence limit is set at 85%.
For the above given example, the accuracy at the lower is
91.6% and hence it is acceptable as the classified map has met
or exceeded the defined accuracy standards.
Commission
No. of pixels
omitted
Total no.
of pixel
% error
No. of pixels
commission
Total no.
of pixels
% error
Heather
32
858
3.7
9.4
920
10.2
Water
878
878
Forest 1
190
910
20.9
21
741
2.8
Forest 2
56
934
6.0
189
1067
17.7
Bare soil
61
621
9.8
29
589
4.9
Pasture
220
0.5
07
226
3.1
( p ' )(q) 50
p p ' 1.96
n
n
Commission
Points
correct
%
correct
95%
confidence
limits
%
correct
95%
confidence
limit
Heather
826
858
96.3
95.0 97.6
920
89.8
87.8 91.8
Water
878
878
100.0
99.4 100.0
878
100.0
99.4 100.0
Forest 1
720
910
79.1
76.4 81.8
741
97.2
95.9 98.5
Forest 2
878
934
94.0
92.4 95.6
1067
82.3
80.0 84.6
Bare soil
560
621
90.2
87.8 92.6
589
95.1
93.3 96.7
Pasture
219
220
99.5
98.3 100.0
226
96.9
94.4 99.1
Class
Using 85% as the criteria, it is seen that Forest 1 has failed the test as both the
upper and lower limits of accuracy at 95% confidence limit is less than 85%.
Similarly, when errors of commission are evaluated, it is found that Forest 2 fails to
meet the criteria. This is also evident from Table .
It can be seen that 183 pixels have not been identified as Forest 1, and has been
identified as Forest 2 class. It implies that even though the overall classification
accuracy has exceeded the defined criterion of acceptability, the training samples
of classes Forest 1 and Forest 2 have not been able to provide the correct
information to the classification process, and hence the training samples have to
be collected with caution.
KAPPA COEFFICIENT
In the above procedure for determining accuracy of
classification, it is highly dependent upon the training samples
used for classification and assessment of classification
accuracy.
In order to assess the agreement between two maps, Kappa
(), which is a measure of the difference between, observed
agreement between two maps (as reported by overall
accuracy) and the agreement that might be contributed solely
by chance matching of two maps.
It attempts to provide a measure of agreement that is adjusted
for chance and is expressed as follows:
Observed Expected
K
1 Expected
789360
753324
635778
915486
505362
193908
858
807760
770884
650598
936826
517142
198428
878
837200
798980
674310
970970
535990
205660
910
859280
820052
692094
996578
550126
211084
934
571320
545238
460161
662607
365769
140346
621
202400
193160
163020
234740
129580
49720
220
920
878
741
1007
589
226
KAPPA COEFFICIENT
Total of diagonal element
= 3046621
= 28866023
= 0.126
0.923 0.126
K
1 0.126
= 0.904
The value of K = 0.904 means that the classification has
achieved an accuracy that is 90% better than would be
expected from random assignment of pixels to classes.
THANK
YOU