Full text: XVIIth ISPRS Congress (Part B5)

  
The image displacement vector d = pa — p4 is called 
optical flow vector and describes the projection of the 
observation point displacement P..1)— P(x) onto the 
image plane. When assuming a linear dependency of 
the surface texture between |, and l? and a brightness 
constancy constraint between frame k and k+1 it is 
possible to predict l2 from I4 and its corresponding 
image intensity gradients and hence to estimate d 
from the measurable difference l» — l4. 
T 
vn (3. 5 d 
y 
Ox 
lo is measured at position of p, at frame k+1, where 
l4 is taken from image position p, at frame k. When 
approximating the spatial derivatives as finite differ- 
ences the optical flow vector d = (dy, dy)" can be 
predicted from the image gradients g = (9x, gy)’ and 
the temporal image intensity difference Alpi z]lo-H 
between frame k«1 and k at p, in Eq. (5): 
Ali = g-d = gd. +g d, 
= 84 * (Pax — Dix) + By * Pay — P1y) 
(4) 
(5) 
In Eq. (5) d is related to intensity differences. Substi- 
tuting the perspective projection of P/x) and P(x41) for 
p, and po in Eq. (5) yields a direct geometric to photo- 
metric transform that relates the spatial movement of 
P between frame k and k+1 to temporal intensity 
changes in the image sequence at p1. 
Al = f- g ; Pax = Pa 
" s Poux Pay 
(6) 
+ f. gy Peay = Pay 
y Puri Puy. 
As long as the motion between Py) and P(k+1) can be 
expressed parametrically by a linear equation and 
assuming that the initial position of Py is known, 
P(k,1)can be substituted and the Eq. (6) canbe solved 
through evaluation of the spatial and temporal intensi- 
ty differences at p4. Essentially every surface point 
can be taken as an observation point. It is, however, 
useful to restrict the number of observation points to 
those carrying relevant information. Eq. (6) uses 
image intensity as well as the spatial image gradients. 
Image areas with zero gradient can not be used for 
parameter estimation. A lower bound to the image 
gradient is additionally introduced to account for cam- 
era measurement noise. It is therefore necessary to 
impose a minimum gradient threshold when selecting 
observation points as described by [Hótter, 1988]. For 
each observation point its initial 3D position on the 
object surface, image intensity and spatial image gra- 
dients are recorded. During the analysis each obser- 
vation point is projected and the intensity difference 
to the real image is evaluated. 
Direct estimation of 3D object motion With the 
proposed approach, rigid 3D object motion can be 
432 
estimated directly from the image sequence when the 
object shape is known. Assume an observation point 
with a known position Py that moves in space and is 
observed by a camera C in Fig. 4. The motion is 
governed by the general motion equation (3). Assum- 
ing that rotation between successive images is small, 
the rotation matrix [Rc] can be linearized to Eq. (7). 
à RAR, 
R] = |R 1 -R, 
-R,R, 1 (7) 
rotation matrix [Rg] is linearized to [R'] 
when setting sin® = o and cos - 1 
When substituting [R’] into Eq. (3) a linearized version 
for the general motion equation is found. P(k+1) is 
expressed in explicit form in Eq. (8): 
(k+1)x 
Pası) = |[Pa+yy] = 
Paix 
(k)x — (Pay — GR; + (Paz — G2)R, * T, 
(Pax — GOR, + Py — (Par — GIR, + T, 
— (Page — GIRy + (Po, - Gj)R, + Poy, TT, 
The parameters to be estimated are the translation T 
and the rotation R. When substituting P(x..1) from Eq. 
(8) in Eq. (6) and linearizing the resulting non linear 
equation through Taylor expansion for R and T at R - 
T « 0, the linearized equation for a single observation 
point Py is computed as 
Alpi f: gx/ Pz : Ty 
f:gy/ Pz Ty 
f-( Pxgx + Pygy) / Pz? Tz 
f: [ Pxgx(Py — Gy) + Pygy(Py — Gy) 
Pzgy(Pz — Gz) ]/ P,2 - Ry 
f-[ Pygy(Px— Gx) + Pxgx(Px - G3) 
P;g«(Pz - Gz) ]/ PZ? - Ry 
f: [gx(Py - Gy) - gy(Px- Gg ]/Pz Rz 
(8) 
+ 
+ 4 "| 
with Pay = (Px Py, PT. (9) 
i imation Atleast six 
distinctive observation points that lead to six linear 
independent equations are needed to solve forthe six 
motion parameters R and T. In real imaging situations 
the measurements of the spatial and temporal deriva- 
tives are noisy and some of the observation points 
selected may be linear dependent of each other. To 
cope with those conditions an over constrained set of 
equations is established and a linear regression is 
carried out using least squares fit. All observation 
points of one object are evaluated. It is important to 
note that we do not measure optical flow locally and 
then try to combine the flow field. Instead all observa- 
tion points of a rigid surface are used to solve for R 
and T. To account for the linearizing, the estimation is 
iterated. The position P of each observation point is 
initially determined by object shape and position. An
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.