Full text: Proceedings of the Symposium on Global and Environmental Monitoring (Part 1)

793 
of the resolution of the imagery. Since the swath width 
is limited, it will frequently be necessary to have several 
overlapping flight lines to obtain the necessary coverage. 
The separate flight lines of imagery will then have to be 
tied to each other and to the existing ground control in 
order to build up a mosaic image of the desired area. 
To put the data volume question into perspective, the 
general experience at CCRS has been that each flight 
line generally contains 8 bands of imagery with 20000 
to 30000 scan lines of 1024 pixels each which requires 
approximately 200 Megabytes of disk storage. Thus for a 
project requiring more than 3 or 4 flight lines, a system 
disk storage capacity of 1 Gigabyte or more very quickly 
becomes a necessity if one wishes to have quick access to 
both raw and processed imagery. 
2 Geometric Processing 
The basic concept behind the geometric correction and 
resampling processing is to compute the exact location 
of each input pixel in the output coordinate system and 
then to resample the imagery onto a uniform grid. Before 
this may be attempted, it is necessary to have accurate 
knowledge of the time history of the sensor motion. Al 
though in theory it should be possible to solve for the mo 
tion of the sensor based entirely on ground control points 
and stereo imagery, the accumulated flight experience at 
CCRS for several different types of aircraft has indicated 
that the aircraft motion contains sufficient high frequency 
components to render the concept impractical. Thus the 
acquisition and recording of auxiliary position and atti 
tude data for use in post-processing of the imagery has 
become the chosen method at CCRS (Gibson, 1986). The 
position and attitude data are recorded from an inertial 
navigation system simultaneously with the imagery data 
in a manner which maintains an exact time correlation 
between the two data sets. Although the inertial sys 
tem parameters contain very good relative information 
(i.e. from sample to sample), they are nevertheless cor 
rupted with low frequency errors in the form of position 
offsets and drifts, and an angular misalignment between 
the inertial coordinate frame and the imager coordinate 
frame. Therefore, before the inertial system data may be 
used to compute the position of the imagery pixels, it is 
necessary to compute and remove these inherent errors. 
This is accomplished by identifying control points in the 
imagery for which the ground coordinates are known and 
then adjusting the navigation system data to achieve a 
least-squares fit of the imagery to the control points. 
2.1 Motion Data 
The motion data recorded in the aircraft consists of veloc 
ity, position and attitude data as derived from the iner 
tial navigation system. In addition, aircraft flying height 
data is also recorded from a high accuracy barometric al 
timeter. At the present time, a project is underway to 
incorporate Global Positioning System (GPS) data with 
the existing inertial system data sets in order to reduce 
or possibly eliminate the requirement for ground control 
point measurements. 
2.2 Photogrammetric Solution 
The algorithm for adjusting the navigation system data is 
based on rigorous photogrammetric principles and incor 
porates the standard Collinearity and Coplanarity condi 
tions (Slama, 1980); the former for utilising ground con 
trol points and the latter for tying multiple flight lines 
together using features of unknown location common to 
overlapping flight lines. The condition equations contain 
non-linear terms involving the attitude angles. A least- 
squares solution to the condition equations may be ob 
tained however, by deriving a linear approximation to 
the equations and following an iterative procedure which 
updates the estimated parameters in each step until no 
further improvement is possible. This approach is well 
known for conventional photo-mapping projects. The al 
gorithm was modified to handle the MEIS situation which 
does not have a central perspective point like a camera 
and also to accommodate the time sampled position and 
attitude measurement data from the inertial navigation 
system. The low frequency random errors in the iner 
tial navigation system data are represented in the system 
error model as low order polynomials and subsequently 
removed as part of the adjustment process. 
The following is a brief overview of the derivation of the 
photogrammetric correction algorithm. Refer to Figure 1 
for illustrations of the following vector definitions: 
• The sensor flight path as a function of time consists 
of a position vector p(t) and an attitude vector a(<), 
both specified with respect to the Geocentric Carte 
sian Coordinate frame X, Y, Z. The attitude angles 
specify the rotation components between the Geo 
centric Cartesian coordinate frame and the sensor 
coordinate frame. 
• The sensor position and attitude at the instant in 
which a specific ground feature is viewed in the 
forward channel are pj and a, respectively, that is 
Pi = vi 1 }) and = «(*>)• 
• The pointing vector for any specific pixel in the for 
ward channel is represented in the sensor coordinate 
frame by the vector m, . 
• The pointing vector (i.e. where the sensor pixel is 
looking for any specific pixel in the forward channel 
is represented in the Geocentric Cartesian coordinate 
frame by the vector xj and is related to mj by the 
equation Xj — Cjirij, where Cj is a direction co 
sine matrix derived from the attitude angle vector 
Cj = C(d}). 
• In a similar manner, the equivalent quantities for the 
aft view are obtained by substituting the subscript k 
in place of j in the above definitions.
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.