Full text: Papers accepted on the basis of peer-reviewed full manuscripts (Part A)

In: Paparoditis N., Pierrot-Deseilligny M.. Mallet C.. Tournaire O. (Eds), IAPRS. Vol. XXXVIII. Part ЗА - Saint-Mandé, France. September 1-3, 2010 
4. METHODOLOGY 
The combination process was implemented in several stages as 
follow: 
4.1 Filtering of lidar point clouds 
First the original lidar point clouds were filtered to separate on- 
terrain points from points falling onto natural and human made 
objects. A filtering technique based on a linear first-order 
equation which describes a tilted plane surface has been used 
(Salah et al.. 2009). Data from both the first and the last pulse 
echoes w’ere used in order to obtain denser terrain data and 
hence a more accurate filtering process. After that, the filtered 
lidar points were converted into an image DTM. and the DSM 
was generated from the original lidar point clouds. Then, the 
nDSM was generated by subtracting the DTM from the DSM. 
Finally, a height threshold of 3m was applied to the nDSM to 
eliminating other objects such as cars to ensure that they are not 
included in the final classified image. 
4.2 Generation of Attributes 
Our experiments were carried out characterizing each pixel by a 
32-element feature vector which comprises: 25 generated 
attributes, 3 image bands (R. G and B), intensity image, DTM, 
DSM and nDSM. The 25 attributes include those derived from 
the Grey-Level Co-occurrence Matrix (GLCM), Normalized 
Difference Vegetation Indices (NDV1), slope and the 
polymorphic texture strength based on the Forstner operator 
(Fbrstner and Gulch, 1987). The NDV1 values for the UNSW, 
Bathurst and Fairfield test areas were derived from the red 
image and the lidar reflectance values, since the radiation 
emitted by the lidars is in the 1R wavelengths. The resolutions 
of the lidar reflectance data for these study areas are lower than 
that for the images, and this may impact on the ability to detect 
vegetation. Since the images derived for the Memmingen 
dataset include an IR channel, the NDV1 was derived from the 
image data only. The attributes were calculated for pixels as 
input data for the three classifiers. Table 3 shows the attributes 
and the images for which they have been derived. These 
attributes have been selected to be uncorrelated based on the 
problem of correlation between feature attributes. All the 
presented attributes were used for every test area. A detailed 
description of the filtering and generation of attributes process 
can be found in Salah et al. (2009). 
attribute 
Red 
Band 
Green 
Band 
Blue 
Band 
Intensity 
DSM 
nDSM 
PTS 
V 
V 
V 
V 
V 
V 
HMGT 
V 
V 
V 
V 
V 
V 
Mean 
V 
V 
V 
V 
V 
V 
entropy 
V 
V 
V 
V 
V 
Slope 
x 
X 
X 
X 
X 
V 
Table 3. The full set of the possible attributes from aerial 
images and lidar data. V and x indicate whether or not 
the attribute has been generated for the image. PTS 
refers to polymorphic texture strength; HMGT refers 
to GLCM/homogeneity; Mean refers to GLCM/ 
Mean; entropy refers to GLCM/ entropy. 
4.3 Land Cover Classification 
In this work, we have used the Self-Organizing Map (SOM), 
Classification Trees (CTs), and Support Vector Machines 
(SVMs) classifiers to estimate the class memberships required 
for the combination process. 
Support Vector Machines (SVMs) 
SVMs are based on the principles of statistical learning theory 
(Vapnik. 1979). SVMs delineate two classes by fitting an 
optimal separating hyperplane (OSH) to those training samples 
that describe the edges of the class distribution. As a 
consequence they generalize w^ell and often outperform other 
algorithms in terms of classification accuracies. Furthermore, 
the misclassification errors are minimized by maximizing the 
margin between the data points and the decision boundary. 
Since the One-Against-One (1A1) technique usually results in a 
larger number of binary SVMs and then in subsequently 
intensive computations, the One-Against-All (1AA) technique 
was used to solve for the binary classification problem that 
exists with the SVMs and to handle the multi-class problems in 
aerial and lidar data. The Gaussian radial basis function (RBF) 
kernel has been used, since it has proved to be effective with 
reasonable processing times in remote sensing applications. 
Two parameters should be specified while using RBF kernels: 
• C, the penalty parameter that controls the trade-off 
between the maximization of the margin between the 
training data vectors and the decision boundary plus the 
penalization of training errors 
• y, the width of the kernel function. 
In order to estimate these values and to avoid making 
exhaustive parameter searches by approximations or heuristics, 
we used a grid-search on C and y using a 10-fold cross- 
validation. The original output of a SVM represents the 
distances of each pixel to the optimal separating hyperplane, 
referred to as rule images. All positive (+1) and negative (-1) 
votes for a specific class were summed and the final class 
membership of a certain pixel was derived by a simple majority 
voting. 
Self-Organizing Map Classifier (SOM) 
The SOM undertakes both unsupervised and supervised 
classification of imagery using Kohonen’s SOM neural network 
(Kohonen, 2001). SOM requires no assumption regarding the 
statistical distribution of the input pattern classes and has two 
important properties: the ability to learn from input data; and to 
generalize and predict unseen patterns based on the data source, 
rather than on any particular a priori model. In this work (Salah 
et al., 2009), the SOM has 32 input neurons which are: 25 
generated attributes, 3 image bands (R, G and B), intensity 
image, DTM, DSM and nDSM. The output layer of an SOM 
was organized as a 15 x 15 array of neurons as an output for the 
SOM (255 neurons). This number was selected because, as 
recommended by Hugo et al. (2006), small networks result in 
some unrepresented classes in the final labelled network, while 
large networks lead to an improvement in the overall 
classification accuracy. Initial synaptic weights between the 
output and input neurons were randomly assigned (0-1). In the 
output of the SOM, each pixel is associated with a degree of 
membership for a certain class. 
Classification Trees (CTs) 
The theory of Classification trees (CTs) (also called decision 
trees) was developed by Breiman et al. (1984). A CT is a non- 
parametric univariate technique built through a process known 
as binary recursive partitioning. This is an iterative procedure in 
which a heterogeneous set of training data consisting of
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.