Full text: Proceedings, XXth congress (Part 5)

USING 3-D VIRTUAL REALITY MODELS FOR IMAGE ORIENTATION 
IN MOBILE COMPUTING 
Charalampos Georgiadis, Anthony Stefanidis, Peggy Agouris 
Dept. of Spatial Information Science and Engineering 
348 Boardman Hall 
University of Maine, USA 
{harris, tony, peggy } @spatial.maine.edu 
Commission V, WG V/5 
KEY WORDS: Close Range, Motion Imagery, Orientation, Virtual Models, Modelling 
ABSTRACT: 
Advances in sensor technology and computing capabilities and modalities are revolutionizing close-range image collection and 
analysis for geospatial applications. These advances create the need for new ways of handling and processing video datasets at quasi 
real time rates. In this paper we present an innovative two step orientation technique for ground level motion imagery using a 3 
dimensional virtual model as control information. In the first step few select anchor frames are orientated precisely via an image 
orientation-through-queries approach. In the second step, intermediate frames are orientated relatively to these anchor frames 
through an innovative analysis of building facade variations in them. Combined, these two steps comprise a complete approach to 
motion imagery orientation using a VR as control information. 
1. INTRODUCTION 
During the last years we are experiencing great advances in 
sensor technology and wireless communications. That progress 
created new data collection schemes, in which users can collect 
data roaming a scene with a GPS enabled digital camera or 
camcorder. These collection schemes create a vast amount of 
data that has to be stored and processed. As a result new 
techniques have to be developed that allow for fast processing 
of motion imagery, either offline or in a quasi real time manner. 
In the context of this paper we use the term motion imagery to 
refer to imagery collected at video rates, or even as select 
frames captured a few seconds (or even minutes) apart, using 
either a video or a still camera. Examples of such datasets 
include imagery collected by hand-held cameras captured while 
roaming an urban environment, or imagery collected by a 
network of fixed sensors (e.g. surveillance cameras) monitoring 
a scene. The computing capabilities of mobile devices are also 
improving with the advances in technology, creating the 
opportunity of location based computing. In order to be able to 
acquire modeling information from such data we have to use 
some kind of control information, the traditional techniques of 
acquiring control points are still time consuming. In addition 
research advances in the development of 3 dimensional virtual 
models of large scale complex urban scenes have resulted in the 
creation of impressive and complex VR models. These 
advances provide opportunities for the integration of such 
models in motion imagery analysis. The accuracy and 
complexity of these models provide an excellent use as control 
information for motion image processing. 
In this paper we present an approach for the recovery of sensor 
orientation and position information using a two-step 
procedure. We focus on the use of motion imagery datasets (at 
quasi-video rates) captured by sensors roaming an urban scene. 
In such a scenario the orientation variations among successive 
frames are small. The first step of our approach entails the 
comparison of object configurations depicted in a frame to 
corresponding configurations identified in the VR model. This 
provides excellent approximate values that are then refined 
(using precise matching) to provide the orientation parameters 
for that frame. This is a process that can be performed for few 
anchor frames (e.g. every few minutes) to provide accurate 
orientation information at these instances. For subsequent 
frames we estimate their orientation by determining their 
variation from the nearest anchor frames. To do so we use 2-D 
transformations of objects depicted in these frames, and 
orientation differences between the frames computed using 
vanishing points to translate these transformation parameters to 
orientation variations. Combined, these two steps comprise a 
novel process of progressive orientation recovery that meets the 
computational requirements of mobile mapping applications. 
Most of the known techniques for computing relative 
orientation need points in different planes. In [Simon and 
Berger, 2002] a similar approach is presented in which they 
estimate the orientation of an image using the previous image of 
known orientation and a planar homography between these two 
images. In another approach [Chia et al, 2002], compute the 
relative orientation based on one or two reference frames, 
exploiting epipolar geometry and using recursive methods. Our 
approach tries to solve this problem using points from only one 
plane, and using linear methods. Furthermore the information 
acquired during the orientation process can be used to update 
the existing virtual model. Such update procedure includes 
change detection both in geometric and radiometric content, in 
the existing objects and detection of new objects or deletion of 
objects in the model. In our paper we present the approach for 
the orientation estimation in the intermediate frames. 
The paper is organized as follows. In section 2 we present an 
overview of our navigation-through-virtual models approach. In 
Section 3 we present the indices we use for comparing the 
intermediate frames with anchor frames for orientation 
recovery. Experimental results in section 4 demonstrate the 
performance of our approach and conclude with future work 
plans in section 5. 
   
   
  
   
   
   
   
  
   
  
   
  
  
   
  
  
  
  
  
   
   
  
   
   
  
  
   
  
  
   
   
   
  
  
   
   
  
  
     
   
  
   
   
    
     
   
   
   
    
    
   
  
  
  
   
   
   
   
  
  
   
   
   
   
    
   
   
  
   
     
Internation 
UE Te 
2. MC 
We assum 
scene that 
database. S 
GPS sensc 
position in 
pose and uj 
Our approa 
-Uthe:! 
to detern 
select an 
— the 
remainin 
Thus we pr 
parameters 
corrections 
orientation 
|. Anchor 
intervals (i 
determined 
Anchor Fri 
  
As we can 
MPEG coi 
frames in 
frames, anc 
the frames 
between c 
philosophy 
position in 
anchor fra 
recovered b 
size of obje 
In figure 2 
approach. 
corresponds 
intermediat 
frame orier 
presented ir 
we should 1 
image quer 
Classic ima 
database ba 
WC use ima
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.