Full text: Proceedings, XXth congress (Part 3)

nbul 2004 
> general 
'onitoring 
ymatic ID 
> analysis 
' used to 
cation of 
cenes are 
>sent and 
mation to 
Azuma et 
curacy or 
nbination 
tracking 
GPS and 
e existing 
lay of a 
ipplied to 
of fixed 
tion. The 
is one of 
2000). It 
lings, so 
important 
specific 
rientation 
niques is 
al., 1996) 
extracted 
n from a 
ed reality 
ors of a 
lected by 
indmarks 
ronment. 
ment an 
ining the 
idmarks). 
d exists — 
; various 
d objects 
integrate 
support 
' devices 
utational 
ces and 
example 
useful to 
ronment. 
nt exists 
ng them. 
pjects as 
igmented 
s to be 
; leads to 
and the 
proaches 
:quential 
nage the 
d images 
International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol XXXV, Part B3. Istanbul 2004 
by image-to-image matching (e.g. based on straight line 
features) to get the accurate position and orientation for the real 
world images taken by the camera later. We do not use such 
comparison methods but a geometric 3D model of the 
environment, providing also 3D information about objects. 
   
pm... 4 
Figure 2. 3D city model (centre of the city of Stuttgart) 
The basic information to provide orientation and navigation by 
image analysis is the availability of a suitable geometric “world 
model”. As example a (geo-referenced) 3D city model 
represents such a world model, see Figure 2. Nowadays 3D city 
models for large areas are easy to provide, as in the meantime 
automatic methods for generation of such models are available 
(Brenner, 2000; Wolf, 1999). For example the City Surveying 
Office of Stuttgart offers such a 3D model of the centre of the 
city for sale. 
2.2.1 Feature extraction: As the principle of our method is 
to co-register image and model data using prominent features in 
the real world and the world model, suitable features must be 
identified in the image, by image analysis techniques. The main 
objects of interest within an urban environment are human made 
objects, e.g. buildings. If we have a closer look to these objects 
so the prominent features of them are texture and edges. The 
edges of human made objects often can be approximated by 
straight lines, therefore we decided to select edges as prominent 
features. Also objects in the 3D city model are modelled by 
straight lines. 
For edge extraction and straight line approximation different 
operators are available, e.g. Burns-Algorithm (Burns, 1986), but 
also the Hough-Transform (Hough, 1962) is a suitable 
algorithm to extract straight lines. Using these algorithms we 
can prepare the collected data (images) and extract the 
necessary feature information. In Figure 3 the result of these 
process is displayed and the extracted straight lines are marked 
in red. 
  
  
   
    
Ji Nebel 9 RE 2 any aJ Vies 
B ORTI DU pee 
ES EE 
— SO 
  
AS | 
a 
  
  
  
    
RT S 
mm 
    
Figure 3. Feature extraction and line detection 
905 
  
On the one hand side the base information for our approach is 
provided by the straight line extraction and on the other hand 
side by the world model (73D city model). The problem that 
occurs now is to select candidate objects within the world 
model for further processing. Additional information provided 
by simultaneous collection of image data as well as rough 
orientation and position information can help to identify one or 
a set of candidate objects. By using our prototype for direct 
collection of exterior orientation parameters we are able to solve 
this problem. The prototype integrates an image sensor (CCD 
camera), an orientation module for goniometry and a GPS 
receiver, which provides rough position information. 
   
    
azimuth 
angle 
vertex 
angle 
CCD »€ 
camera 
  
Figure 4. Prototype sensor 
To extract required data for further processing we have to query 
the model for objects, which are contained in the viewing 
frustum of the camera and select the visible objects (see Fig. 5). 
The data collected by the prototype sensory can help us to 
extract these model data by an integrated processing of sensor 
and model data. 
  
.&, 
"A. 
Figure 5. Exttaction of visible objects 
2.2.2 
World Model) to image registration problem the features of the 
selected and extracted 3D object model and the features 
extracted in the 2D image must be assigned. This is the task of 
determining straight lines in image and object space that are 
belonging together. Automatic registration of 3D models to 
Feature Matching: To solve the model (Augmented 
images is a fundamental and open problem. It is difficult 
because it comprises two coupled problems: the correspondence 
and the pose problem. Each sub problem is easy to solve only if 
the other has been solved first. A classic approach for solving 
these coupled problem is the hypothesize and test approach. 
First a small set of image feature to model feature 
correspondences are hypothesized, e.g. using sensor data 
collected by the prototype sensor to minimize the search space. 
Based on the result the object's pose is computed and the model 
is back projected into the image. The pose is accepted if the 
original and back projected image are sufficiently similar, 
otherwise a new hypothesis is formed and the process is 
repeated. As example the method of Beveridge and Riseman 
(Beveridge&Riseman, 1995) uses a random start local search 
with a hybrid pose estimation algorithm employing both full- 
perspective and weak-perspective camera models. To simplify 
the equations often linear affine approximations (weak- 
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.