Full text: Close-range imaging, long-range vision

OOo OO O°O”OO9\O©°OO©°)© OO OO OO OO OO OL OO CC 
Ob 00 © 00°” C°0202020 © 090°0°© 900°» 9 02009 © 0 Q 
Q'ODOOOGXooocdocoocoooodooocccuo 
nages, it has been 
e and therefore fast 
'enerally sufficient, 
idated blobs being 
subsequent bundle 
ages on the other 
typically identified 
cess. A shape test 
tion process for the 
st suitable for low- 
the following three 
ow-quality images: 
on via a best-fitting 
nerally require only 
lunder’ pixels from 
because of inherent 
thm. The procedure 
ziation of intensities 
unders those pixels 
mean minus a given 
rocess is needed to 
on for the ellipse fit 
to remove regions 
hot-spots. Whereas 
mogenous reflection 
nsity distribution of 
ally quite inhomo- 
es the pixels of such 
try as dark blunder 
region to the point 
ze testing. 
ellipse to the target 
region. The centre 
ed for the ellipse fit 
been described, for 
| Luhmann (2000). 
a best-fitting ellipse 
or large targets (say 
ions invariably pass 
ry point distribution 
estimation of ellipse 
| of departures of the 
à 
/ 
2.4 Sub-Pixel Centroiding 
Once a target region has passed the validation process its exact 
centre needs to be determined. Though various centroiding and 
template matching methods can be found in the literature (e.g. 
Atkinson, 1996; Luhmann, 2000; Shortis et al. 1994), the 
intensity-weighted centroiding approach is most often adopted 
in automated VM because it is simple and very fast to compute: 
n m x. 
kN : 
Xo | _ i=l j=l Yij 
n m 
Yo 
Ej 
i=l j=1 
@) 
Here, x; and y; are row and column coordinates of pixels within 
the target blob, g; is the corresponding grey value, and (x,, y,) 
are the final centroid coordinates. The main issue of concern in 
any high-precision  centroiding method is background 
thresholding. The threshold has to be chosen as low as possible, 
so it removes all background noise without deleting valid edge 
pixels of the target region. The target edge pixels preserve the 
important geometric information for sub-pixel centroiding, 
which in VM can reach accuracies to 2-3% of the pixel size. For 
today’s large area CCD cameras, this can translate to an image 
coordinate accuracy of close to 0.2 um. 
3. EO DEVICE DETECTION 
Effective automatic detection of the EO device is necessary for 
initial spatial resection of the images forming the VM network, 
which is in turn critical to a successful solution of the image 
point correspondence problem using epipolar geometry. 
Multiple strategies for automatic recognition of target groupings 
within an image can be designed. Theoretically, recognition 
strategies for coded targets could also be employed for the EO 
device, but the main difference between coded targets and an 
EO device is that the points on the EO device have known XYZ 
coordinates within an arbitrary datum, which initially defines 
the reference coordinate system of the full VM network. One 
reason that EO devices are designed to be noticeably larger than 
coded targets is to guarantee sufficient geometric strength for 
the calculation of EO parameters for all images via either 
closed-form resection or, less commonly, relative orientation 
Whereas EO devices in commercial VM systems such as 
V-STARS (GSI, 2002) use recognisable fixed patterns of targets, 
a more flexible recognition and initial EO concept has been 
developed for Australis, where the user has the flexibility to 
define their own EO device under minimal design constraints. 
Figure 4 shows an example EO device suited to Australis. This 
design incorporates one basic constraint, namely that all points 
must be enclosed within a bounding shape, a circle in this case 
(a rectangle, square or triangle would also suffice). To achieve 
better approximations of orientation parameters, one of the five 
targets is usually non-coplanar with the others. Use of the EO 
device requires no more than the user specifying the XYZ 
coordinates of the EO target points. No labelling is necessary so 
long as the targets are asymmetrically distributed, since their 
relative positions are decoded in the recognition process. 
The EO detection procedure can be subdivided into three stages: 
device detection, target recognition and pattern decoding, and 
closed-form spatial resection. For the adopted procedure, the 
pattern decoding and resection stages are closely interrelated. At 
the initial image scanning stage, the primary recognition 
criterion for the EO device is the closed boundary, which need 
not be of a specific shape. The segmentation process treats the 
boundary in a similar way to other targets when forming an 
appropriate segmentation matrix, an extract from which is 
shown in Figure 5 for an EO-device boundary section. During 
the EO validation process it is checked whether the segmented 
region has a ‘hole’, ie an enclosed region within the boundary, 
the test being performed only if the enclosed area contains a 
specified minimum number of pixels. 
  
Figure 4. An example EO device for Australis. 
Following the image scanning, the segmentation matrix for a 
detected enclosed region is tested to establish whether it 
contains others targets. If the number of such detected targets 
matches that for an EO device, the region is flagged as an EO- 
device candidate. The design for Australis can account for the 
detection of multiple EO devices in an image, though this 
feature has not been applied in practical applications so far. 
Once the correct enclosed region and all EO-device targets are 
detected and measured, the pattern decoding and closed-form 
resection are performed. 
0 0 0 0 0 0 0 28 28 28 28 28 28 2B 2B 28 28 28 28 
0 0 0 0 0 28 28 28 20 20 20 20 28 28 28 28 28 28 28 
0 0 0 28 28 20 28 28 28 28 28 28 26 28 28 28 28 28 28 
0 28 28 28 26 28 28 28 28 28 28 28 28 28 26 28 28 28 28 
B 28 28 28 28 28 28 28 20 20 28 28 28 28 28 28 28 20 26 
8 28 28 20 20 28 28 28 28 28 28 28 28 20 28 28 28 28 28 
B 28 28 28 28 28 28 28 28 28 26 28 28 28 28 28 28 28 0 
28 28 28 28 28 28 28 28 28 28 28 28 28 28 20 D 0 0 0 
28 28 28 28 28 28 28 28 28 28 28 28 28 28 0 0 0 0 
28 28 28 28 28 28 28 28 28 20 28 28 28 28 0 0 0 00 
28 28 2B 28 28 28 28 28 28 29 28 28 28 0 0 0 0 
28 28 28 28 28 28 28 28 28 28 28 0 0 0 0 
28 28 28 28 28 28 28 282820 0 0 0 0 0 © 
28 28 28 28 28 28 28 28 0 0 00 
E 2028 20 28 28 28 28 0 0 0 0 00 
28 28 28 28 28 2 
  
  
  
  
  
  
  
  
  
  
   
9 
    
   
0 00 
0 
0 
0 
o 
0 
900 
00 
00 
00 
00 
0 0 
0 
0 
506000000 
ecocooo 
0 
0 
0 
0 
0 
0 
0 
0 
9 
00 
9 0 
0 
...........>.>{ 
oooococooooocoo 
9 90 
0 0 0 474747 0 
0 0 4747474747 47 0 
0 
0 
9 
0 
o 
0 
0 
0 
0 
9 
0 
o 
0 
° 
0 
o 
ooooocoooooocococs 
0 
0 
0 
0 
0 
0 
0 
0 
0 
0 
0 0 47 47 47 47 47 47 0 
D 0 47 47 47 47 47 47 0 
0 0 4747 47 47 47 47 0 
0 004747474747 0; 
0000009000; 
e 
ooooooocooocococooooo 
o6o0oo0ooooooooooooo 
o 
0 
0 
[i 
0 
0 
0 
90 
0 
0 
o 
0 
0 
0 
0 
ooooócooon 
50500000 
o 
0 
0 
o 
  
Figure 5. Image of EO device with part of segmentation matrix. 
To determine which image point corresponds to the correct 
object point, a simple but rigorous strategy is employed. An 
algorithm tries all permutations of possible resections to find the 
most likely solution. If the geometric arrangement of EO points 
is favourable (an asymmetric pattern) the sought after 
permutation is the one with the lowest RMS value of image 
coordinate residuals in the resection. This algorithm has proven 
to be very robust in practical applications, even in cases where 
sensor calibration parameters (primarily lens distortion and 
focal length) are only approximately known, or where the lens 
distortion is assumed to be zero and a very course initial 
principal distance is adopted. 
63 — 
 
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.