Full text: XVIIth ISPRS Congress (Part B3)

  
  
the measure of distances of pixels from class centres, 
the sample that best represent a specific land feature 
is the one which best define the centre of such class. 
In conventional methodologies samples are either a) 
arbitrarily selected by visual inspection of a video 
display of imagery data (supervised classification) or 
b) automatically selected by implementing some form 
of clustering technique on raw data (unsupervised 
classification). Both of these techniques could possibly 
work only if color distinctions between populations of 
terrestrial objects would be sharp and no color overlap 
would exist between neighbouring classes, which is 
certainly not the case in remote sensing. In fact, in 
the majority of cases only a few radiance levels 
separate a class from the next. Even a skillful image 
analyst would find very difficult, if not impossible, to 
outline visually on a video screen samples that will 
define the actual location of class centres, so that valid 
statistical discriminating functions can be applied. 
Moreover, if variance values are also computed from 
these samples, as normally done with the maximum 
likelihood decision rule, the wrong assumption is also 
made that the sample variance, which in the majority 
of cases depends mostly on local factors (e.g. the 
presence of discontinuities in the canopy of crops), 
does somehow reflect the actual variance of an entire 
class of objects. As to the application of clustering 
techniques, the presence of *mixture" pixels together 
with the color overlap existing among various classes 
makes it very difficult to find valid histogram peaks, 
i.e. peaks that do represent class centres. 
3. GEOCLASS APPROACH 
3.1 Image Segmentation 
Let us look now at the solutions to these problems 
provided by the GEOCLASS approach. First of all, it 
was found necessary to isolate object populations from 
each other through a segmentation process based on 
the extraction of boundaries that delimit regions of 
homogeneous color. This process is done in stages 
and is illustrated by Figure 1. The upper left quadrant 
depicts a 4.5x3.5 kms agricultural test site located in 
Manitoba, Canada. This site .vas used for an in-depth 
evaluation of the capabilities of the TM sensor to 
discriminate a variety of field crops: potatoes, flax, 
cereals, peas, rapeseed (canola), etc. (Steffensen and 
Mack, 1986). This enhanced portion of a TM scene 
gathered on July 1,1984 was produced using band 4: 
red, band 7: green, band 2: blue. Fallow and potatoes 
appear as green tones, cereals as red tones, peas and 
canola as pink tones, woodlands as dark red tones. 
The upper right quadrant shows the results of 
extracting gradient values for each pixel. This is done 
by measuring a non-directional gradient value 
occurring within a 3x3 cell centered on each pixel. 
This process is applied band by band. If a pixel is 
surrounded by pixels of the same value in a specific 
band it will have a zero gradient value in that 
particular band. The higher the difference between 
the center pixel and its eight neighbours, the stronger 
the gradient value. The end result of this automated 
computation is a subimage in which dark areas are 
areas of color homogeneity and bright lines mark the 
color changes occurring within the scene. It can be 
seen on figure 1 that high gradient values affect a few 
pixels at the edge of each agricultural field and that 
different gradient values occur in different bands, 
depending on the crop types involved. Yellow lines 
(high gradient values in bands 4 and 7) mark the 
boundaries between exposed soil and vegetation, while 
purple lines (high gradient values in bands 4 and 2) 
mark boundaries between different green crops. To 
further understand how the upper right quadrant of 
figure 1 is related to the upper left quadrant, one can 
904 
focus his attention on the triangular feature 
appearing on the upper left quadrant. This is an 
abandoned airport. Since the runways are 
represented by more than one pixel in width the 
gradient algorithm is capable of isolating a black low- 
gradient pixel having on both sides bright high- 
gradient pixels (see right side). Now, by applying a 
ridge-edge extraction algorithm to the upper right 
quadrant we obtain the subimage illustrated in the 
lower left quadrant, where precise boundary lines are 
defined. This process is an iterative process whereby 
the image analyst chooses the proper thresholds for 
achieving satisfying results. The final step in this 
segmentation process is an automated filtering of the 
boundary lines of the lower left quadrant from the 
upper left quadrant to obtain the lower right quadrant. 
A zero value in all bands is assigned to the filtered out 
pixels resulting in the black areas shown on the lower 
right quadrant. Notice that the parcels resulting from 
the segmentation process do follow in the large 
majority of cases in shape and size the different field 
crops and that only rarely individual fields are divided 
in multiple sub-units. 
3.2 Training 
Assuming that each agglomerate of pixels within a 
parcel portrays a single object, which is normally the 
case, the segmentation process provides us with the 
object population needed for classification. A unique 
identification number is assigned to each parcel, 
allowing for the average color of parcels to be 
computed in each band. Then these vector data are 
displayed as a color scattergram on the video screen of 
an image analysis system. The analyst can either 
assign to the points in the scattergram identical colors 
to those of the corresponding parcels in the image, or 
any transformed color (e.g. ratios between bands). In 
this way the analyst is provided with the capability to 
a) identify visually the location of class centres by 
grouping together scattergram points having similar 
color; and b) define valid training samples, by 
selecting parcels located around class centres as 
being representative of each class. In summary, the 
basic advantage of the GEOCLASS approach in 
training is that it allows the identification of valid 
locations for class centres, which is the paramount 
factor for a successful classification. 
3.3 Classification 
In the last phase of the classification process a 
classifier is applied to extend the classification from 
the training samples to the entire image. The 
conventional approach is to carry out this final stage 
as if the structural context of each pixel would be ot no 
significance. In other words, if an image is 
scrambled, or if we would change arbitrarily the 
relative position of the pixels, there would be no 
impact on the classification results. However, 
contextual considerations can be quite helpful in 
finding the correct classification for pixels not having 
a distinct signature and for boundary pixels. 
Boundary pixels should not be processed similarly to 
“pure” pixels. In a conventional system class 
validation is done purely on theoretical grounds. 
Valid classes are those for which the computation of 
certain statistical parameters (e.g. confusion matrix) 
indicates no class overlap in feature space. In this 
case, a high level of accuracy is expected in the final 
classification results. However, even samples that are 
not overlapping in feature space may lead to an 
unsatisfactory classification in terms of thematic 
accuracy, if these samples are not located close to 
class centres. Even relatively small changes in 
selecting training samples can significantly change 
EN CD | ch E sou Ee obe4 pie Ey (ho B PS PS LI a py es AN Mm mde pu m uk Um. Iun Gc o d 1. qu. 707577 
rt pad Co uri Pal 
CN ONTO
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.