Full text: Close-range imaging, long-range vision

there is no texture 
also not an option. 
detect the overall 
an extracting single 
ape is more robust 
ion by trees, cars, 
nto the scene. The 
ition for its overall 
CAD model of the 
according to the 
n how to select the 
Fritsch 2001). The 
ict the silhouette of 
on is then detected 
  
  
| for a given exte- 
low. 
nsform (GHT) to 
"Work for both the 
sional shapes in 
tect the shape no 
illy even scaled in 
sedom, since the 
litionally the GHT 
viation, which is 
1g is only a coarse 
in the image. 
rm (Hough, 1962) 
es or circles in an 
n of the Hough 
‘as a function of a 
on of that concept 
a simple analytic 
e edge magnitude 
ase, which has to 
so-called R-table 
s selected usually 
stance vector r of 
vith respect to its 
n the R-table. Of 
re can be several 
is performed, all 
1. For each edge 
y in the R-table, 
which possibly holds several, vectors ri. These vectors 
correspond to positions in the image, which receive a vote. The 
position in the image receiving the most votes at the end is 
selected as the position of the shape in the image. 
If the orientation of the object is allowed to vary, as is the case 
in our application, a separate R-table has to be computed for 
each discrete rotation angle. The same is true for scaling. Thus 
the formation of the R-tables in our case is quite complex and 
computationally expensive. 
P 
Figure 3: Properties of a single edge pixel P as recorded in the 
framework of the Generalized Hough Transform. 
For our implementation we made use of the HALCON image 
processing environment, which provides a shape detection 
mechanism based on the GHT (Ulrich, et. al. 2001). In order to 
compensate for the computational costs of large R-tables, this 
operator includes several modifications to the original GHT. 
For example it uses a hierarchical strategy generating image 
pyramids to reduce the size of the tables. By transferring 
approximation values to the next pyramid level the search space 
is drastically reduced. 
3. ORIENTATION 
The result of the GHT is a two-dimensional similarity 
transformation consisting of translation, rotation and scale. This 
transformation corrects for the misalignment of the shape to its 
actual appearance in the image. In other words for each three- 
dimensional point of the original CAD model rendered to the 
two dimensional image we can use the transformation to correct 
its location in the image. 
3.1 Spatial Resection 
This gives us the possibility to create point wise pseudo 
observations. We can select an arbitrary point from the CAD 
model of the building and project it onto the image using the 
calibration data of the camera and the initial approximated 
orientation. Then we use the two-dimensional transformation 
from the GHT to move the image point to its correct location. 
From a minimal configuration of three points we can compute 
an improved exterior orientation of the camera by 
photogrammetric spatial resection. 
Several alternatives for a closed form solution to the resection 
problem are given in literature. We follow the approach 
suggested by (Fischler & Bolles 1981). Named the “Perspective 
4 Point Problem” their algorithm solves for the three unknown 
coordinates of the projection center when the coordinates of 
four control points lying in a common plane are given. Because 
the control points are all located on a common plane the 
mapping in-between image- and object points is a simple plane- 
to-plane transformation T. The location of the projection center 
can be extracted from this transformation T when the principal 
distance of the camera is known. For a detailed description of 
the formulas please refer to the original publication. 
To complete the solution of the resection problem we also need 
the orientation of the camera in addition to its location. (Kraus 
1996) gives the solution for determining the orientation angles 
when the coordinates of the projection center are already 
known. The algorithm makes use of only three of the four 
points. 
In principle, the complete process, extraction of building 
silhouette, improvement of image coordinates by GHT and 
spatial resection can be repeated iteratively in order to avoid 
errors resulting from the initial orientation data. Nevertheless, 
for our application the differences between the projected wire- 
frame and the image were mainly caused by errors within the 
available model due to measurement errors or generalization 
effects. Subsequent iterations did not enhance the result 
significantly. 
4. EXPERIMENTS AND RESULTS 
A series of real-world examples was chosen from our database, 
to investigate both the feasibility of our approach and the 
accuracy obtained from photogrammetric processing. Our 
investigations are based on a dataset of the city of Stuttgart 
provided by the City Surveying Office of Stuttgart. This data 
was collected by manual photogrammetric stereo measurement 
from images at 1:10000 scale (Wolf 1999). For data collection 
the public Automated Real Estate Map (ALK) was additionally 
used. Thus, a horizontal accuracy in the centimeter level as well 
as a large amount of detail is available. 
  
Figure 4: Prototype of the mobile photogrammetry device con- 
sisting of a camera, a compass and a tilt sensor. The GPS is 
mounted separately. The laser unit was not used in this project. 
495 
  
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.