Full text: Proceedings; XXI International Congress for Photogrammetry and Remote Sensing (Part B1-3)

The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences. Vol. XXXVII. Part Bl. Beijing 2008 
1060 
one needs to check whether the estimated internal 
characteristics of these cameras from a calibration session 
remain stable over time. The question that arises from such a 
need is what are the tools and standards that can be used to 
evaluate the stability of a given camera? 
This paper will thus focus on these issues in relation to amateur 
digital cameras. First, an automated method for an indoor 
camera calibration procedure is introduced. The main objective 
of such a procedure is to provide mapping companies using 
these cameras with a simple calibration procedure that requires 
an easy-to-establish test field. More specifically, a test field that 
is comprised from a set of linear features and few point targets 
will be used to carry out the camera calibration procedure. The 
determination of the interior orientation parameters will be 
based on the observed deviations from straightness in the image 
space linear features as well as the measured distances between 
the point targets. In addition to having a simplified calibration 
test field, we will outline an automated procedure for the 
extraction of the linear features and point targets from the 
captured imagery. The simplified test field and the automated 
extraction of the linear features and point targets are the key 
factors for enabling the data providers to effectively carry out 
the calibration procedure, which is general enough to handle 
any amateur digital camera. The paper will then discuss the 
concept of how to evaluate camera stability, which will be 
followed by the introduction of a set of tools for its evaluation. 
The stability analysis tools will be based on quantitative 
evaluation of the degree of similarity between the reconstructed 
bundles from temporal calibration sessions. The bundle 
similarity is used since the camera calibration procedure aims at 
generating a bundle that is as similar as possible to the incident 
bundle onto the camera at the moment of exposure. 
Following the calibration and stability analysis discussions, the 
paper will deal with several related questions as follows: 1) 
How to develop meaningful standards for evaluating the 
outcome from the calibration procedure, 2) How to develop 
meaningful standards for evaluating the stability of the involved 
camera, 3) Is there a flexibility in choosing the stability analysis 
tool, which is commensurate with the geo-referencing 
procedure to be implemented for this camera, and 4) Can the 
stability analysis be used for evaluating the equivalency of 
different distortion models. These questions will be discussed in 
turn, and some experiment results from datasets captured by 
two amateur small format digital cameras are presented. 
2. CAMERA CALIBRATION 
Deriving accurate 3D measurements from imagery is contingent 
on precise knowledge of the internal camera characteristics. 
These characteristics, which are usually known as the interior 
orientation parameters (IOP), are derived through the process of 
camera calibration, in which the coordinates of the principal 
point, camera constant and distortion parameters are determined. 
The calibration process is well defined for traditional analogue 
cameras, but the case of digital cameras is much more complex 
due to the wide spectrum of designs for digital cameras. It has 
thus become more practical for camera manufacturers and/or 
users to perform their own calibrations when dealing with 
digital cameras. In essence, the burden of the camera calibration 
has been shifted into the hands of the data providers. There has 
thus become an obvious need for the development of standards 
and procedures for simple and effective digital camera 
calibration. 
Control information is required such that the IOP may be 
estimated through a bundle adjustment procedure. This control 
information is often in the form of specifically marked ground 
targets, whose positions have been precisely determined 
through surveying techniques. Establishing and maintaining this 
form of test field can be quite costly, which might limit the 
potential users of these cameras. The need for more low cost 
and efficient calibration techniques was addressed by Habib and 
Morgan (2003), where the use of linear features in camera 
calibration was proposed as a promising alternative. Their 
approach incorporated the knowledge that in the absence of 
distortion, object space lines are imaged as straight lines in the 
image space. Since then, other studies have been done by the 
Digital Photogrammetry Research Group (DPRG) at the 
University of Calgary, in collaboration with the British 
Columbia Base Mapping and Geomatic Services (BMGS), to 
confirm that the use and inclusion of line features in calibration 
can yield comparable results to the traditional point features. In 
order to include straight lines in the bundle adjustment 
procedure, two main issues must be addressed. The first is to 
determine the most convenient model for representing straight 
lines in the object and image space, and secondly, to determine 
how the perspective relationship between corresponding image 
and object space lines is to be established. In this research, two 
points were used to represent the object space straight-line. 
These end points are measured in one or two images in which 
the line appears, and the relationship between theses points and 
the corresponding object space points is modelled by the 
collinearity equations. In addition to the use of the line 
endpoints, intermediate points are measured along the image 
lines, which enable continuous modelling of distortion along the 
linear feature. The incorporation of the intermediate points into 
the adjustment procedure is done via a mathematical constraint 
(Habib, 2006a). It should be noted, however, that in order to 
determine the principal distance and the perspective centre 
coordinates of the utilized camera, distances between some 
point targets must be measured and used as additional 
constraints in the bundle adjustment procedure. 
To simplify the often lengthy procedure of manual image 
coordinate measurement, an automated procedure is introduced 
for the extraction of point targets and line features. The steps 
involved in the procedure are described in detail in Habib 
(2006a) and are briefly outlined in the following sub-section. 
2.1 Automated Extraction of Point and Line Features 
The acquired colour imagery is reduced to intensity images, and 
these intensity images are then binarized. A template of the 
target is constructed, and the defined template is used to 
compute a correlation image to indicate the most probable 
locations of the targets. The correlation image maps the 
correlation values (0 to ±1) to gray values (0 to 255). Peaks in 
the correlation image are automatically identified and are 
interpreted to be the locations of signalized targets. Once the 
automated extraction of point features is completed, the focus is 
shifted to the extraction of linear features. The acquired 
imagery is resampled to reduce its size, and then an edge 
detection operator is applied. Straight lines are identified using 
the Hough transform (Hough, 1962), and the line end points are 
extracted. These endpoints are then used to define a search 
space for the intermediate points along the lines. Once the point 
and linear features have been extracted through this automated 
procedure, they are incorporated into the bundle adjustment, 
according to the method outlined in Section 2, to determine the 
camera IOP.
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.