Full text: Technical Commission III (B3)

XXIX-B3, 2012 
09. Automatic Road 
n Classifier Fusion, 
pp. 1-6 
-based simplification 
Graphics 24 (2), pp. 
2005. Mesh saliency, 
lings of SIGGRAPH 
ialitative comparison 
oftware ^ packages. 
mendedPubl/Vitaly/ 
zical characterisation 
esis, University of 
04. The Automatic 
, University of New 
Australia, 
load?doi=10.1.1.133 
). Automatic Road 
Classifier Fusion , 
Joint Event, 
arnumber-5137739 
L., 2010.Quanfu B. 
AR Data Based on 
| Congress on Image 
2133. 
CLASSIFICATION BY USING MULTISPECTRAL POINT CLOUD DATA 
Chen-Ting Liao ^ *, Hao-Hsiung Huang * 
* Department of Land Economics, National Chengchi University, 
64, Sec. 2 Zhinan Rd., 11605 Taipei, Taiwan — 99257006@nccu.edu.tw, hhh@nccu.edu.tw 
Commission III, WG III/2 
KEY WORDS: Classification, Image Matching, Close Range Photogrammetry, Infrared, Point Cloud 
ABSTRACT: 
Remote sensing images are generally recorded in two-dimensional format containing multispectral information. Also, the semantic 
information is clearly visualized, which ground features can be better recognized and classified via supervised or unsupervised 
classification methods easily. Nevertheless, the shortcomings of multispectral images are highly depending on light conditions, and 
classification results lack of three-dimensional semantic information. On the other hand, LiDAR has become a main technology for 
acquiring high accuracy point cloud data. The advantages of LiDAR are high data acquisition rate, independent of light conditions 
and can directly produce three-dimensional coordinates. However, comparing with multispectral images, the disadvantage is 
multispectral information shortage, which remains a challenge in ground feature classification through massive point cloud data. 
Consequently, by combining the advantages of both LiDAR and multispectral images, point cloud data with three-dimensional 
coordinates and multispectral information can produce a integrate solution for point cloud classification. Therefore, this research 
acquires visible light and near infrared images, via close range photogrammetry, by matching images automatically through free 
online service for multispectral point cloud generation. Then, one can use three-dimensional affine coordinate transformation to 
compare the data increment. At last, the given threshold of height and color information is set as threshold in classification. 
1. INTRODUCTION 
Passive remote sensing systems record electromagnetic energy 
reflected or emitted from the surface as two- dimensional 
multispectral images. The general used bands are blue 
(0.45~0.52um), green (0.52~0.60um), red (0.63~0.69um), Near 
Infrared (0.70~1.3um), Middle Infrared (1.3~3um) and thermal 
Infrared (3~14um). Due to ground features have their own 
characteristic in different spectrum, while classifying through 
Multispectral Images, generally, higher divergence between 
bands, may lead to higher classification accuracy. Therefore, 
one can interpret ground features effectively by collecting 
multispectral images, for example, healthy vegetation reflects 
massive near infrared light, and water body absorbs near 
infrared light, so one can use near infrared light with other 
bands for recognizing vegetation and water body. 
LIDAR is an active remote sensing system, which can acquire 
ground feature point cloud data through laser scanning 
technique; this allows remote sensing data development toward 
three-dimensional space. Point cloud data includes three 
dimensional coordinates, intensity and other abundance spatial 
information, which contains much more potential to interpret 
ground features than two-dimensional image does. In general, 
LIDAR scans ground features by single band laser light, for 
instance, green lasers at 0.532um has water penetration ability, 
and vegetation has high sensitive to near infrared laser light 
region in 1.04um to 1.06um (Jensen, 2007). Generally, point 
cloud data is acquired only through single band laser light, and 
lack of multispectral information. 
Consequently, this research uses close-range photogrammetry 
method to collect visible light and near infrared images, and 
" Corresponding author. 
135 
chooses free online service — Photosynth, which is provided by 
Microsoft, as automatically image matching technique for point 
cloud generation. After exporting the point cloud data, one can 
use three-dimensional affine coordinate transformation to merge 
multispectral point cloud and visible light point cloud data, as a 
check for the accuracy and precision for multispectral point 
cloud data. Comparing with point cloud data generated by using 
visible images, increment of multispectral point cloud data 
acquired by adding near infrared images were then evaluated. 
Thereafter, the multispectral point clouds for ground feature 
were classified. The results of classification have been analysed, 
for understanding whether the point clouds generated by 
multispectral information have good potential in classification. 
2. BACKGROUND 
Ground features have diffuse reflectance properties respectively. 
Understanding of the spectral reflectance of ground features can 
assist in distinguishing and recognizing diffuse ground features. 
Generally, collecting visible light image can only acquire 
spectral reflectance from 0.4um to 0.7um by collecting other 
band, e.g. near infrared light, the spectrum beyond visible light 
can be obtained, which can interpret ground feature effectively. 
By matching multispectral images through free online service, 
such as Photosynth, one can get point cloud data from image 
collected in close range; via three-dimensional coordinate 
transformation, combining it with visible light point cloud data, 
therefore one can compare and analyze the benefit of 
multispectral images on increasing point cloud data and 
classification assisting ability. 
The following sections will introduce the advantage in ground 
feature interpretation by adding near infrared, brief introduction 
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.