International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, Vol XXXV, Part B3. Istanbul 2004
3.3 Experimental Results
For each image, it is exactly key that the projected line is
extracted from this image. In the experiment, the curve is
extracted out by the image processing skill. According to the
characteristic of the image and the regulation of the coordinates
movement of the extracted curve, the section of the curve on
the body of the convenient cup is separated from the whole
curve. For example, the experiment processes the first image in
the Figure 2. The experimental results are shown as two images
of the Figure 3.
|
Figure 3. The whole curve and the section of the curve
The points on the section of the extracted curve are available
for the next calculation. The coordinates of these points are
expressed by the unit of pixel. There are separately 284 points,
294 points, 297 points and 249 points in the first image, the
second image, the third image and the fourth image of Figure 2.
Every point on the four sections of the curve has its own
homologous point on the line of the projected slide through the
homologous epipolar lines. So the space coordinates of four
sets of space points can be computed out by the space forward
intersection. The two end points of each set of the space points
is picked out and their coordinates are used in the whole
adjustment. The results of the whole adjustment are shown in
the Table 1.
X, (mm) R, (mm) | H (mm)
Yo (mm)
-0.0954 5.7235 23.9147 80.7226
Table 1. The results of the whole adjustment
(x) Vp) is the coordinates of the centre of circle. R, is the
radius of circle of the mouth of the convenient cup and R, is
the radius of circle of the bottom of the convenient cup. /7 is
the height of the convenient cup.
Known the coordinates of the centre axis, there is a radius of
circle in every horizontal level according to the coordinates of
the points of the extracted curves. The whole model of the
convenient cup is made up of many levels of the circles. The
view of the final model is shown in the Figure 4.
(c) From lean (d) From top
Figure 4. The model of the target cup
4. CONCLUSIONS
4.1 Achievements of the Paper
The approach proposed in the paper is confirmed to be proper
and applicable from the results of the experimental data. This
approach only requires a digital camera, an ordinary slide
projector and a rotating platform. These equipments are easy to
be ready for the applications.
The paper also deduces the detail algorithm of this approach
which is understandable and relatively simple. Using the
algorithm provided, the correct model results can be gotten
from the image data. To sum up, the approach with this
algorithm is effective and effectual.
According to the size of the object, the distance of the slide
projector from the rotating platform and the point density of the
slide can be changed and adjusted, so that this approach is
suitable for many kinds of applications. Moreover, it is hardly
affected by the space factor or time factor. The approach
provided by the paper is flexible and practical.
4.2 Limitations and Future Work
Because of the tight time and the restricted experimental
environment, there are a few limitations appearing in this
approach.
€ According to the size of the rotating platform and the
intensity of the light of the slide projector, the distance
from the slide projector to the rotating platform is not too
International Ar
ee 61 X
far and the
platform. €
applicable
e The reflec
object is c
material o:
clear or 1
extraction
reduplicate
need be go
® The exper
real textur
target obje
add to the
® The exper
solid of ro
to be con:
complex s
To these limitati
as following.
® The featur
detailed ac
size, its sl
should be
procedure.
® When the
feature pre
with no 1
functioned
object. Th
by the in
relative im
e Ihe3Dr
divided in
object is r
affiliation
the main b
To sum up, the
on the basis of t
paper.
Thanks for the :
the Doctoral Pr
2001048601 1).
Deren "Li, IS
University Pres:
Jianging Zhang
technique for t
SPIE - The Ir
Beijing, China,