Full text: Proceedings, XXth congress (Part 1)

    
  
  
  
   
   
    
   
    
  
   
  
  
   
  
   
    
   
   
  
   
   
  
  
  
   
   
  
    
    
     
    
   
   
  
    
   
  
   
       
   
    
   
  
  
     
     
   
   
1 2004 
s from 
cheme 
arch in 
it lines 
'enient 
ince it 
On the 
juence 
eature. 
' space 
| cause 
ig ina 
ints in 
under 
ie the 
te that 
| other 
| in all 
3, the 
  
  
Cr 
D. 
asured 
onding 
olished 
will be 
points 
igh a 
n this 
to any 
hin the 
ind the 
Figure 
al. 
(D) 
ig the 
Lt space 
to the 
is the 
nediate 
noted 
gina 
rdinate 
image 
ntation 
(OP) 
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol XXXV, Part B1. Istanbul 2004 
  
including distortion parameters, as well as the ground 
coordinates of the points defining the object space line. Such a 
constraint does not introduce any new parameters and can be 
written for all intermediate points along the line in the imagery. 
The number of constraints is equal to the number of 
intermediate points measured along the image line. 
AY 
“a 
X 
o^ o? 
   
   
| | 
LA -Z. 
x. =x — distortions x 
OK) , — distortions y 
—c 
X, e X, 
5e r-5 
12, -Z, 
“1 “0 
Figure 2: Perspective transformation between image and object 
space straight lines. 
As a special case of the above procedure, the treatment of 
control linear features (with known object coordinates of its end 
points) will be slightly different. The control line will provide 
the end points in the object space; hence, these end points need 
not be measured in any of the images and no collinearity 
equations will be written for any of the control lines. 
Subsequently, image space linear features are represented only 
by a group of intermediate points measured in all images. 
LIDAR straight-line features 
The growing acceptance of LIDAR as an efficient data 
acquisition system by researchers in the photogrammetric 
community led to a number of studies aiming at pre-processing 
LIDAR data. The purpose of such studies ranges from simple 
primitive detection and extraction to more complicated tasks 
such as segmentation and perceptual organization (Maas and 
Vosselman, 1999: Csathó et al., 1999; Lee and Schenk, 2001; 
Filin, 2002). 
In this paper, LIDAR straight line features will be used as a 
source of control for photogrammetric models. To extract such 
lines, suspected planar patches in a LIDAR dataset are 
manually identified with the help of corresponding optical 
imagery, Figure 3. The selected patches are then checked using 
a least squares adjustment to determine whether they are planar 
or not and to remove blunders. Finally, neighbouring planar 
patches with different orientation are intersected to determine 
the end points along object space discontinuities between the 
patches under consideration. 
In another approach to simplify the extraction process, intensity 
and range data recorded by the LIDAR system are utilized for 
direct measurement of linear features. Raw range and intensity 
data are first interpolated to a uniform grid using identical 
interpolation method and parameters. Linear features previously 
extracted from photogrammetry are then identified on the 
intensity image from which planimetric coordinates of line ends 
are measured while observing height readings from the range 
image, Figure 4. It is worth mentioning that the interpolation 
method and applied parameters have a visible effect on the 
accuracy of this approach. 
   
(a) (b) 
Figure 3: Manually identified planar patches within the LIDAR 
data (a) guided by the corresponding optical image (b). 
Sa 
  
FEPEIRE 
(a) à 
  
Figure 4: Manually measuring planimetric coordinates from 
intensity image (a) and height from range image (b). 
2.2 Approach 2: Using LIDAR lines in the absolute 
orientation of photogrammetric model 
This approach starts with generating a photogrammetric model 
through a photogrammetric triangulation using an arbitrary 
datum without knowledge of any. control information. The 
datum is achieved through fixing 7 coordinates of three well- 
distributed points in the bundle adjustment procedure. The next 
step is determining the elements of the absolute orientation 
parameters to align this photogrammetric model to the LIDAR 
reference frame using conjugate straight line segments. Both 
photogrammetric and LIDAR line segments are represented by 
their end points. These end points are not required to be 
conjugate. In this paper, a 3D similarity transformation is used, 
Equation 2. 
X. Xr X 
Y deze +S Ame LL: 0 
Z4 Zr Zz 
Where S is the scale factor, (X4 Y4 Z4). is the translation vector 
between the origins of the photogrammetric and LIDAR 
coordinate systems, R is the 3D orthogonal rotation matrix, (X, 
Y, Z,). are the point coordinates in one dataset, while (X4 YA 
Z4). are the point coordinates in the other. 
Referring to Figure 5, the two points describing the line 
segment from the photogrammetric model undergo a 3-D 
similarity transformation onto the line segment AB from the 
LIDAR dataset. The objective here is to introduce the necessary 
constraints to describe the fact that the model segment (12) 
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.