Full text: Proceedings; XXI International Congress for Photogrammetry and Remote Sensing (Part B4-3)

AN ACCURACY ASSESSMENT MEASURE FOR OBJECT BASED IMAGE 
SEGMENTATION 
nent s and 2/3 of a 
' ni P accuracy 
reflected 
in 
» the 
are measured 
Prevement 
in the 
ges from different 
ger ) " S1( k lap h/h 
ction. 
ad analog images 
acies with small 
ma g e $ also allow 
)f 5 Pixels. The 
; al accuracy of up 
and buildings. In 
se (approximately 
lot well represent 
lings or other man 
elusion is that the 
lore information, 
ore comfort in the 
it loss of the relief 
ieight accuracy, it 
itions in the final 
n the horizontal 
al component can 
:ts. 
¡ages in the paper 
enerated from a 
ire known. The 
the LSF function 
lion of the system, 
tion, using USAF 
Iso be applied to 
The results then 
r of the DMC, in 
Vrbiol, R.. Boned 
¡z, M. À., Palma, 
fanon process of 
“High-Resolution 
nnover. 
seeker, S;* 
F ,f DMC. to 
I, Paris. 
ie selection 
»rial triangulati° n 
Nicholas Clinton 3 ’*, Ashley Holt 3 , Li Yan b , Peng Gong 3C . 
“Department of Environmental Science, Policy and Management. University of California, Berkeley CA 94720-3114 
b Intemational Institute for Earth System Science. Nanjing University, China. 210093 - nclinton@nature.berkeley.edu 
c State Key Laboratory of Remote Sensing Science, Jointly Sponsored by Institute of Remote Sensing Applications, Chinese 
Academy of Sciences, and Beijing Normal University, 3 Datun Road, Chaoyang District, Beijing 100101 
KEY WORDS: Segmentation, ASTRO, BerkeleylmageSeg, Ecognition, Definiens 
ABSTRACT: 
Traditional approaches to accuracy assessment are inadequate for object oriented image processing. We tested some measures to assess 
the accuracy of object based image segmentation in a supervised context. The measures quantify the extent to which objects in the 
segmentation match training objects in terms of over-segmentation, under-segmentation, and distance to a perfect match. Using high 
resolution digital aerial photographs over an urban setup, we obtained segmentation results for a variety of parameter combinations using 
two software packages: eCognition and ASTRO. We compute the accuracy measures using three types of objects: vehicles, trees and 
buildings. The measures were used to compare the software, identify ideal parameter combinations, and identify objects that each 
software is better at extracting from the images. The measures are shown to be an intuitive, useful technique for consistency checking 
different segmentation results and assessing segmentation accuracies among a large set of disparate segmentation results. 
INTRODUCTION 
In object based image processing, the first step is generally a 
segmentation of the image of interest. A wide variety of 
segmentation results may be obtained through different parameter 
combinations or different segmentation software. Prior to 
classification or even to training of a suitable classifier, one of the 
segmentation results must be chosen. In this paper, we describe 
well defined measures that can be used in the identification of a 
“best” segmentation and the “best” objects within that 
segmentation for training a classifier. These measures are 
applicable in the supervised setting only, and “bestness” is 
therefore relative to a set of pre-defined training objects (assumed 
polygons) over the image of interest. 
Assume first that the landscape of interest is a finite population of 
objects (Bian 2007). The spatial information about these objects 
is useful in the ultimate classification of the object (Gong and 
Howarth 1990). It is obvious that exact representation of the 
objects in the segmentation is important, since this shape 
information will eventually be presented to a classifier for the 
identification of a pattern. The accuracy of the classification is 
thus dependent on the accuracy of the shape information 
submitted to the classifier. Measures of the segmentation result 
are therefore relevant to the interpretation and optimization of 
ultimate classification accuracy. The measures we tested are not 
measures of classification accuracy, but are related. Assuming 
that accuracy assessment is conducted with statistical rigor, a 
probability sample will be obtained on the population of objects 
(Stehman and Czaplewski 1998, Stehman 1999). If the population 
is assumed to be represented by the segmentation result and a 
simple random sample is used to generate accuracy statistics, then 
the accuracy of the shapes has been completely ignored! On the 
other hand, if a sample is taken from the landscape directly (e.g. 
human delineated training polygons are used) and compared to the 
segments, then areas of intersection between mapped classes and 
reference classes affect the resultant accuracy. The accuracy of 
the segmentation will thus directly influence the classification 
accuracy, unless classification is performed on object primitives, a 
different problem discussed below. 
There are a large number of methods with which to judge 
segmentations (Zhang 1996). This study is focused on the 
scenario in which a set of training objects is available for a static 
image and segmentation results are to be compared to these pre 
defined training objects. Unlike unsupervised evaluation of 
segmentation results (Levine and Nazif 1985, Ng and Lee 1996, 
Borsotti et al. 1998, Chabrier 2006), spectral aspects (such as 
homogeneity within segment or within class) of the resultant 
segments are not considered and the quality of segments is 
evaluated solely in respect to the shape of training objects. In this 
context, a segmentation result should contain segments that match 
the training objects. For automatically checking this, a simple, 
intuitive measure of polygon matching can be computed. This 
measure relies on the observation that there should be a one-to- 
one correspondence (in area) between human identified objects 
(training objects) and segments. A measure of this 
correspondence was first proposed by Levine and Nazif (1982) 
and demonstrated by Yang et al. (1995). Moller et al. (2007) 
describe a similar measure called Relative Area (RA) which relies 
on the ratio of intersected area to segment and reference object 
area. We present the intuition behind this measure, a refinement 
of its computation, and a case study using high resolution urban 
imagery. 
SEGMENTATION GOODNESS MEASURES
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.