Full text: Technical Commission III (B3)

    
  
   
   
   
  
   
  
   
   
    
  
  
   
   
   
  
  
  
  
  
  
   
   
   
   
   
   
   
   
  
   
  
   
   
   
   
    
e XXXIX-B3, 2012 
b 
c denotes the translation 
^s. For this, it is assumed 
the center between both 
jon 
n calculated for the Ieg- 
zed for detecting special 
Iculating the coarse area 
ng features which should 
ss as they arise from ob- 
hese features have to be 
ickground being relevant 
e representations of sev- 
ibase before starting the 
images contains a tem- 
ne, but from a different 
ice and at a different sea- 
intensity images show a 
e detected SIFT features 
> to the object templates 
1 maximum similarity to 
transformation based on 
template is transformed. 
emplate is then assumed 
lure allows for detecting 
ne as well as for decou- 
the presented approach 
ase of dynamic environ- 
ly known. 
IMAGING SYSTEMS 
1e scene monitoring with 
à future operational sys- 
ces fairly realistically, a 
owever, due to the large 
le system, mounting the 
1 and data storage on an 
impracticable. Hence, 
active multi-view range 
long a rope is used as 
ire 3. The components 
ision] CamCube 2.0) for 
sk for efficiently storing 
pendent power supply. 
ge imaging devices can 
variable multi-view op- 
or divergent data acqui- 
nce of noise effects aris- 
liation in comparison to 
ipath scattering, the uti- 
range accuracy of a few 
expected. Furthermore, 
time-of-flight cameras, 
  
  
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume XXXIX-B3, 2012 
XXII ISPRS Congress, 25 August — 01 September 2012, Melbourne, Australia 
  
Figure 3: PMD[vision] CamCube 2.0 and model of a cable car 
equipped with two range imaging devices. 
depends on the modulation frequency fm, where co denotes the 
speed of light. A modulation frequency of 20 MHz thus corre- 
sponds to à non-ambiguous range of 7.5 m. In order to overcome 
this range measurement restriction, image- or hardware-based un- 
wrapping procedures have been introduced (Jutzi, 2009; Jutzi, 
2012). When dealing with multiple range imaging devices, it also 
has to be taken into account that these may influence each other 
and that interferences are likely to occur. This can be overcome 
by choosing different modulation frequencies. 
4 EXPERIMENTAL RESULTS 
The estimation of the flight trajectory of a sensor platform re- 
quires the definition of a global world coordinate frame. This 
world coordinate frame is assumed to equal the local coordinate 
frame of the sensor platform at the beginning. The local coor- 
dinate frame has a fixed orientation with respect to the sensor 
platform. It is oriented with the X -direction in forward direction 
tangential to the rope, the Y -direction to the right and the Z- 
direction downwards. For evaluating the proposed methodology, 
a successive pairwise registration is performed. The threshold for 
the matching of 2D features is selected as tdes = 0.7. The result- 
ing 2D/2D correspondences are projected into 3D space which 
yields 3D/3D correspondences. Including the weights in the esti- 
mation of the rigid transformation yields position estimates and, 
finally, an estimated trajectory which is shown in Figure 4 in nadir 
view and in Figure 5 from the side. The green and blue points 
describe thinned point clouds captured with both range imaging 
devices and transformed to the global world coordinate frame. 
  
Y [m] 
  
  
  
  
X [m] 
Figure 4: Projection of the estimated trajectory and thinned point 
cloud data onto the X Y -plane. 
A limitation of the experimental setup seems to be the fact that 
no reference values are available for checking the deviation of 
the position estimates from the real positions. However, due to 
the relative orientation of the sensor platform to the rope, the 
projection of the real trajectory onto the X Y -plane should ap- 
proximately be a straight line. Additionally, the length of the real 
trajectory projected onto the ground plane can be estimated from 
aerial images or simply be measured. Here, the distance Aground 
between the projections of the end points onto the ground plane 
has been measured as well as the difference Aaititude between 
maximum and minimum altitude. From the measured values of 
Âground = T m and Acititude = 1.25 m, a total distance of 
   
  
Z [m] 
  
  
  
X [m] 
Figure 5: Projection of the estimated trajectory and thinned point 
cloud data onto the X Z-plane. 
approximately 7.11 m can be assumed. A comparison between 
the start position and the point with the maximum distance on the 
estimated trajectory results in a distance of 6.90 m. As a con- 
sequence, the estimated trajectory can be assumed to be of rela- 
tively high quality. The results for a subsequent object detection 
and segmentation is illustrated for an example frame in Figure 6. 
  
Figure 6: SIFT-based object detection and segmentation: normal- 
ized active intensity image, template and transformed template 
(upper row, from left to right). The corresponding point cloud for 
the area of the transformed template and the sensor position (red 
dot) are shown below. 
5 DISCUSSION 
The presented methodology is well-suited for dynamic environ- 
ments. Instead of considering the whole point clouds, the prob- 
lem of registration is reduced on sparse point clouds of physically 
almost identical 3D points. Due to this fact and the non-iterative 
processing scheme, the proposed algorithm for point cloud reg- 
istration is very fast which is required for monitoring in such 
demanding environments. Although the current Matlab imple- 
mentation is not fully optimized with respect to parallelization of 
tasks, a total time of approximately 1.63 s is required for pre- 
processing, point quality assessment, feature extraction and point 
projection. Further 0.46 s are required for feature matching, cal- 
culation of weights and point cloud registration. This can signif- 
icantly be reduced with a GPU-implementation of SIFT, as the 
calculation of SIFT features already takes approximately 1.54 s. 
Furthermore, the simple estimation of a rigid transformation is 
not sufficient, as used 3D/3D correspondences have the same 
weight, even if the uncertainty of the respective 3D points is very 
high or if outlier correspondences not fitting to the transforma- 
tion have been detected. Hence, a quality measure for 3D/3D 
correspondences has been introduced which is based on the qual- 
ity of the respective 3D points. This quality measure is used for 
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.