ISPRS Commission III, Vol.34, Part 3A ,,Photogrammetric Computer Vision“, Graz, 2002
A à ud amem € m) x ?
d MASS 0
(d) (e) (f)
p* ie
A iiti, ü J % /
| | | 4% /
| | | | 1 48 i
| | /
| | [ | mes. 5 : /
B " | J Sy J
1 ! ag, © 18 J 14 3 4 j
{ A i i
uid i i "B J 45 J
m i 4 f. jp sinet /
EF | - 1 £N / z
a SI
, iii ; e
13 / a
(9) (h) (i)
Fig. 3: Results of fusing aerial images with LIDAR data. Fig. (a) shows the test site of Ocean City and (b) depicts the
sub-area that was used to demonstrate the detailed surface reconstruction (indicated by white box in a). In (c), the results
of segmenting the LIDAR point cloud is shown with a total of 19 planar surface patches. The next step of the perceptually
organized LIDAR point cloud is shown in (d) where adjacent planes are intersected, resulting in the six breaklines, I1 16.
Fig. (e.f) contain the edges of the aerial stereopair obtained by the Canny operator, illustrating the difficulty of image
matching for stereopsis. The next figures (g,h) show more specific edges that are obtained using the current knowledge
about the scene. These edges were matched in object space with the segmented LIDAR surface. The final result of
combining information from aerial and LIDAR is shown in (i). The color code for the region boundaries corresponds to: red:
LIDAR+aerial(stereo); yellow: LIDAR+aerial(mono); magenta: aerial(stereo); blue: LIDAR; green: aerial(mono).
A - 317