Full text: Papers accepted on the basis of peer-reviewed full manuscripts (Part A)

In: Paparoditis N., Pierrot-Deseilligny M.. Mallet C... Tournaire O. (Eds), IAPRS. Vol. XXXVIII. Part ЗА - Saint-Mandé, France. September 1-3, 2010 
154 
(as optional choice). Its interest points are matched with a 
multi-image geometrically constrained LSM (Grim and 
Baltsavias, 1988) based on the collinearity principle and the 
orientation parameters previously computed with SIFT or 
SURF features. Table 3 shows the accuracy improvement using 
the developed approach. 
2.4 Correspondence reduction and uniform distribution 
High resolution images picturing objects with a good texture 
could generate a huge number of image correspondences, which 
are very often not uniformly distributed in the images. 
Therefore, each image can be divided into rectangular cells 
where only the features with the largest multiplicity are held. 
This method demonstrated a significant reduction of the 
correspondences (even more than 100 times), preserving the 
accuracy of the final product. Secondly, as the features are 
generally random-distributed in the images without a uniform 
distribution (Figure 3), the method improves the geometric 
distribution of tie points. The quality of the final results is 
increased in terms of geometric distribution of tie points and in 
a drop of the processing time and computational cost. 
Figure 3. Random and non-uniform distribution of the features 
extracted with SIFT (above) and derived uniform distribution 
with the proposed approach (below). 
2.5 Extension to spherical images 
ATiPE can be also used to extract reliable correspondences 
from spherical images (or panoramas), i.e. a mosaic of 
separated images acquired with a rotating head and stitched 
together. The derivation of metric information from spherical 
images for interactive exploration and realistic 3D modeling is 
indeed receiving great attention due to their high-resolution 
contents, large image field of view, low cost, easiness, rapidity, 
and completeness (Fangi, 2007). But spherical images, if 
unwrapped on a plane, feature different resolutions (width and 
height), scale changes and the impossibility to use a classical 
bundle solution in Cartesian coordinates. In fact, while a 
pinhole image is described by its camera calibration parameters, 
a generic spherical image is only described with its 
circumference C, which corresponds to the image width (in 
pixels) under an angle of 2n. In fact, a spherical image can be 
intended as a unitary sphere S around the perspective centre and 
3D point coordinates x can be expressed in terms of longitude A 
and co-latitude ig as: 
x = [x y z] r = [sin^cosA sin ^ sin A cos^] r (1) 
The relation between a point onto the sphere and its 
corresponding 3D coordinates is x = X / || X ||. Homogenous 
image coordinates m can be mapped onto the sphere by using 
the equi-rectangular projection: 
m = [m ] m 2 l] r =[/?/l Rig l] r (2) 
where R=CI(2n) is the radius of the sphere. 
Given a spherical image, we developed a matching strategy to 
unwrap the sphere onto a local plane. First of all, the median of 
the longitudes p(A) is subtracted from A, obtaining new 
longitudes A* =A - p(A). This allows the projection of the points 
of the sphere onto the plane x = 1 as: 
p = [l p, p y ] T = [l tgA* coty/7cosA*] r (3) 
Here, p 2 and p 3 can be intended as the inhomogeneous image 
coordinates of a new pinhole image. Moreover, the centre of the 
spherical images is also the projection centre of the new pinhole 
image, with the advantage that given two spherical images S 
and S', an outlier can be removed by robustly estimating a 
fundamental matrix. Obviously, this procedure cannot cope 
with large longitude variations. However, the partitioning of the 
spherical images into 4 zones (lad2 <A< (¿+1)7t/2, k = 0,..., 3) 
produces 4 local pinhole images that can be independently 
processed. In addition, this method allows the combined 
matching of spherical and pinhole images. 
The extracted image correspondences are then processed with a 
bundle solution in spherical coordinates to derive the camera 
poses and a sparse 3D geometry of the analyzed scene. 
3. EXPERIMENTS 
3.1 Ordered image sequences 
Figure 4a shows the recovered poses for a sequence matched 
with the SIFT operator and the kd-tree search for comparing its 
descriptors. It took roughly 30’ to orient 33 images (used at 
their original size) acquired with a 10 Megapixel calibrated 
camera. The robust estimation of the epipolar geometry was 
necessary to remove mismatches (e.g. moving objects such as 
people and pigeons). The orientation procedure was carried out 
starting with a relative orientation between the first image pair, 
then progressive resections alternated to triangulations have 
been used to provide approximations for the final and robust 
photogrammetric bundle adjustment. Figure 4b shows the poses 
of 28 images recovered with the SURF operator in 15’, deriving 
a final sparse point cloud of ca 12,000 3D tie points. 
Figure 5 shows the results from the ISPRS Comm. 3 dataset 
“Fountain-K6”. The 25 images have been automatically 
oriented with the SIFT operator in approximately 20’. The 
adjustment with all the extracted features (more than 96,000 
image points) gave a final RMS of 3.6 pm, while the 
application of the point reduction described in subsection 2.4 
(ca 6,000 image points) ended with a RMS of 3.3 pm. 
3.2 Sparse image block 
ATiPE has been also tested on large image blocks acquired with 
an UAV system. Figure 6 shows the orientation results for a 
block of 70 images taken with a calibrated 12 Megapixel 
camera. The global number of image combination is 2,415, 
while the pre- analysis with the visibility map found 507 image 
combinations.
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.