Full text: XVIIth ISPRS Congress (Part B3)

  
The reasoning method just illustrated is called per- 
fect induction. Here, the available information which 
is necessary to support the desired logical inference is 
perfect. 'This means that all the statements only have 
two values for their validities, either true or false, and 
we are able to check exhaustively all the possible com- 
binations. Unfortunately, in many practical problem- 
solving situations, especially in image analysis and 
understanding, the available knowledge is incomplete 
or inexact. The coordinates of the points in 7, for ex- 
ample, may contain measure-errors and only from the 
fact that some points of P do not fulfill y =a =z + b 
we can not come to the conclusion that P does not 
depict a straight line. So, the validity of Y (or ^Y) 
is not binary and not easy to prove if we do not have 
knowledge about measure-errors. In cases like this, we 
need reasoning methods in making just decisions. 
4 Inverse Problems 
Mathematically, the inverse problem can be described 
as follows. Given a mapping f from set A’ into set , 
ie. f : Y! — y. The solution of the inverse problem 
consists in the interpretation of data Y € yy in order 
to recover the original image X € À. This is exactly 
the same goal as that of an inductive inference men- 
tioned above. 
Let us now consider a linear mapping A : A — y. 
The inverse problem is to identify X from the data 
Y: 
AX Y. (1) 
The solvability of this inverse problem could be dis- 
cussed as follows: 
e If A is bijective and A-! is stable one can easily 
get an unique solution X = A-!Y. 
e If A is injective but not surjective, the inverse 
problem is overdetermined and has no solution. 
One can, however, get an unique pseudo solu- 
tion through minimizing the norm of the residual 
IV = lY — AXI. 
e If A is not injective, the inverse problem is under- 
determined and there is an unique pseudo solu- 
tion X — A*Y, where A* is the so called Moore- 
Penrose Inversion, which, unfortunately, is only 
stable if the domain of A is closed in Y. 
So, it is quite clear that the ambivalent non-injective 
inverse problem is practical not solvable through a 
numeric process, as any few errors in Y can destroy 
the solution totally. 
Schematically, there are two reasons for the ill- 
posedness of inverse problems: intrinsic lack of data, 
490 
and observation uncertainties. With additional infor- 
mation, for instance some a priori assumptions on 
model parameters X or an additional data set, many 
such problems can be reformulated into well-posed 
solvable problems. Now, the main question is how to 
integrate a priori knowledge to solve ill-posed inverse 
problems. We need criteria to impose constrains on 
the solution space and a framework to integrate a 
priori knowledge in order to select an unique solution 
(the so called best solution) for given data. Intuiti- 
vely, the best solution exists only in connection with 
criteria which are, of course, strongly task dependent. 
5 The MAP Criteria 
The first criterion which we would introduce in this 
section is the so called Maximum A Posteriori (MAP) 
criterion which is based on probability theory (Geman 
and Geman, 1984 ). It selects as the best solution the 
model parameters X that maximizes the conditional 
probability of X given the data Y: P(X | Y), subject 
to the inverse problem (1). The MAP criterion leads 
to three important estimation methods, namely the 
bayesian estimate method (BE), the maximum like- 
lihood method (ML) and least squares method (LS) 
(cf. Tab. 4), which are widely used in data processing. 
Using Bayes’ theorem gives 
P(X |Y) = P(Y | X)P(X)/P(Y), (2) 
where P(Y | X) is the conditional probability of get- 
ting data Y given the model parameters X, P(X) are 
the prior probability of X. The relation (2) shows how 
the prior probability P(X) changes to the posterior 
probability P(X | Y) asa result of acquiring new infor- 
mation Y. Intuitively, the MAP criterion will choose 
X that maximizes 
P(Y | X)P(X), (3) 
if P(Y) is constant. This is the principle of the 
Bayesian estimation. Further, under the specification 
that the prior probabilities P(X) are all the same, 
ie. P(X) is constant, the MAP criterion leads to 
the simpler maximum likelihood principle of selec- 
ting that X which maximizes P(Y | X). If the ran- 
dom variables to which the data Y refer are normally 
distributed, the maximum likelihood estimation will 
give the same results as the least squares estimation 
which has widely been used in different branches of 
science and engineering for over a century and a half. 
If V is the vector of observational residuals, for which 
E(V) = 0, and which is assumed to be normally dis- 
tributed, and X is the covariance matrix of the distri- 
bution, then we have 
P(Y | X) = P(V) = C - exp |-5v7=="v] ud) 
  
whe 
crit: 
to 1 
miz 
So f 
proi 
posi 
dely 
for | 
deal 
und 
rion 
mew 
the 
beh: 
into 
ter : 
sing 
prox 
and 
whic 
and 
niqu 
6 
The 
hand 
meni 
spac 
ing t 
ver, 
P(X 
The 
a ph; 
ter s] 
a pos 
all pc 
job is 
priat. 
meas:
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.