Full text: Technical Commission III (B3)

   
2.6 Point Cloud Registration 
The spatial relation between two point clouds with n 3D/3D cor- 
respondences X; €» X; with X;, X; € mR? can formally be 
described as 
X;=RX;+t (6) 
where R. € R?*? represents a rotation matrix and t € IR? rep- 
resents a translation vector. A fully automatic estimation of the 
transformation parameters can be derived from minimizing the 
error between the point clouds. Including a weighting w; € R for 
each 3D/3D correspondence X; € X; yields an energy function 
E with 
E — ) w|X; - (RX; t| (7) 
i=1 
for the registration process. For minimizing this energy function 
E, the registration is carried out by estimating the rigid trans- 
formation from all 3D/3D correspondences and the weigths are 
derived from a histogram-based approach. This approach is ini- 
tialized by dividing the interval [0m, 1m] into n, — 100 bins of 
equal size. For all detected correspondences, the calculated qual- 
ity measures c; and oc; assigned to the 3D points X; and X are 
mapped to the respective bins b; and b;. Points with standard 
deviations greater than 1 m are mapped to the last bin. The oc- 
currence of mappings to the different bins is stored in histograms 
h = [h;], 100 9nd h' — [e| a "da Loo Subsequently, cu- 
mulative histograms 
h.= Ec " and hi = ba 
j=1 j=1 
are derived. The entries of the cumulative histograms reach from 
0 to the number n of detected correspondences. As points with 
a low standard deviation are more reliable, they should be as- 
signed a higher weight. For this reason, the inverse cumulative 
histograms 
izl,.., 100 i=1,..., 100 
Beine = ^ TE S " (8) 
juil 
i=l,..., 100 
and 
ÿ=l i=i. 100 
are calculated. Finally, the weight w; of a 3D/3D correspondence 
X; + X; is set to 
w; = minfh. n0(0:), hEinu( 71)! (10) 
where c; and c; are considered as quality measures for the re- 
spective 3D points X; and X;. Estimating the transformation 
parameters can thus be carried out for both range imaging de- 
vices separately. However, as the relative orientation between the 
devices is already known from a priori measurements and both 
devices are running synchronized, the rigid transformation can be 
estimated from the respective correspondences detected by both 
devices between successive frames. Combining information from 
both devices corresponds to extending the field of view and this 
yields more reliable results for the registration process. The ex- 
tension can be expressed by transforming the projected 3D points 
X, which are related to the respective camera coordinate frame 
(superscript c) into the body frame (superscript b) of the sensor 
platform according to 
XR XI +t (11) 
422 
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Volume XXXIX-B3, 2012 
XXII ISPRS Congress, 25 August — 01 September 2012, Melbourne, Australia 
where R? describes the rotation and t^ denotes the translation 
between the respective coordinate frames. For this, it is assumed 
that the origin of the body frame is in the center between both 
range imaging devices. 
2.7 Object Detection and Segmentation 
As 2D SIFT features have already been calculated for the reg- 
istration process, they can also be utilized for detecting special 
objects in the scene. This allows for calculating the coarse area 
of an object and for automatically selecting features which should 
not be included in the registration process as they arise from ob- 
jects which are likely to be dynamic. These features have to be 
treated in a different way as the static background being relevant 
for registration. For this purpose, image representations of sev- 
eral objects have to be stored in a database before starting the 
surveillance application. One of these images contains a tem- 
plate for the object present in the scene, but from a different 
measurement campaign at a different place and at a different sea- 
son. Due to a similar altitude, the active intensity images show a 
very similar appearance. Comparing the detected SIFT features 
of the normalized active intensity image to the object templates 
in the database during the flight yields a maximum similarity to 
the correct template. Defining a spatial transformation based on 
the SIFT locations as control points, the template is transformed. 
The respective area of the transformed template is then assumed 
to cover the detected object. This procedure allows for detecting 
both static and moving objects in the scene as well as for decou- 
pling sensor and object motion. Hence, the presented approach 
for registration also remains reliable in case of dynamic environ- 
ments if representative objects are already known. 
3 ACTIVE MULTI-VIEW RANGE IMAGING SYSTEMS 
The proposed concept focuses on airborne scene monitoring with 
range imaging devices. For simulating a future operational sys- 
tem involving such range imaging devices fairly realistically, a 
scaled test scenario has been set up. However, due to the large 
payload of several kilograms for the whole system, mounting the 
required components for data acquisition and data storage on an 
unmanned aerial vehicle (UAV) still is impracticable. Hence, 
in order to investigate the potentials of active multi-view range 
imaging systems, a cable car moving along a rope is used as 
sensor platform which is shown in Figure 3. The components 
mounted on this platform consist of 
e two range imaging devices (PMD[vision] CamCube 2.0) for 
recording the data, 
e anotebook with a solid state hard disk for efficiently storing 
the recorded data and 
e a 12 V battery with 6.5 Ah for independent power supply. 
As the relative orientation of the two range imaging devices can 
easily be changed, the system allows for variable multi-view op- 
tions with respect to parallel, convergent or divergent data acqui- 
sition geometries. 
However, due to the relatively large influence of noise effects aris- 
ing from the large amount of ambient radiation in comparison to 
the emitted radiation as well as from multipath scattering, the uti- 
lized devices only have a limited absolute range accuracy of a few 
centimeters and noisy point clouds can be expected. Furthermore, 
due to the measurement principle of such time-of-flight cameras, 
the non-ambiguous range R, with 
Hs (12) 
s. 
Im 
N| = 
  
   
   
  
   
   
   
  
  
  
   
    
    
   
   
    
  
  
   
    
    
    
    
   
    
    
    
    
    
     
    
    
   
   
   
     
  
    
    
  
Figure ? 
equippe 
depends 
speed o 
sponds t 
this rang 
wrappin 
2012). \ 
has to b 
and that 
by choo 
The est 
quires t 
world c 
frame o 
dinate f 
platforn 
tangent 
directio 
à succe: 
the mat: 
ing 2D/ 
yields 3 
mation 
finally, 
view ar 
describ 
devices 
Y [m] 
Figure 
cloud d 
A limit 
no refe 
the pos 
the rel: 
project 
proxim 
trajectc 
aerial 1 
betwee 
has bei 
maxim 
A row 
  
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.