Full text: From pixels to sequences

  
  
242 
have implemented the following sequence: f) — c) — a) — b) — d) — e). 
The calculation of each of these features returns a float value between 0.0 and 1.0, where 0.0 means that the 
feature is absent and 1.0 that the feature is present. These values are then used to assess the usability of the 
vertex pair as a grasping opportunity. The result should be consistent with the feature-values, e.g., if one feature 
is 0.0, the result should be 0.0, too. Fuzzy set theory provides several functions for the conjunctive combination 
of evidence, the so-called triangular norms, or T-norms which have the appropriate qualitative effect, among 
them T; (a, b) = min(a, b) or T(a, b) — a x b. Using the latter, our final grasping opportunity measure for one 
vertex pair becomes: 
Ti = Uf fi * UcCi * Vals * Ub; * vadi * veei 
where 7; stands for “ranking of vertex pair i” and v, denotes the relative importance of feature x, with 
x € {a,b,c,d,e, f}. Finally, we select the vertex pair with the highest ranking and the corresponding grasp is 
carried out, i.e., the object concerned by the grasp is removed. 
5 VISION SYSTEM INTEGRATION 
A state diagram of the complete vision system is shown in Figure 3. A short description of the tasks performed 
in the different states follows: 
- Start: The vision system and the robot are initialized. 
- Image acquisition: The two depth images are taken and merged (as described in section 3.1 and 3.2) 
- Build world representation: The input image is segmented and "objects" are assigned to each con- 
nected surface region (see section 3.3). Several predicates are assigned to each object, such as "isolated" 
or “on top of”. All objects are also marked "graspable". 
- Finding grasping opportunities: As described in section 4, this state consists of an “object selection 
sub-state” and a state which calculates the actual grasping points. 
- Grasp object: The grasp is carried out by the robot. Independently of the success of the grasp, a new 
image is taken. 
- Mark object *non-graspable": If there are no suitable grasping opportunities for an object, it is 
marked “non-graspable” and thus is not considered anymore by the "object selection state". 
- Push highest object: If there are still objects in the scene but all of them are marked "non-graspable",. 
the vision system has failed. The reason might be a high degree of occlusions, reflecting or highly textured 
objects etc. The system now tries to change the spatial constellation of the scene. This is done in the 
hope that the deadlock is removed. 
© 
Wait for Image acquisition 
next scene 
: Build world 
Push highest representation 
object 
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
Figure 3: State diagram of the 
Scene empty | Object selection | complete vision system 
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
all objects : " Mark object = 
marked "non-graspable" object selected non-graspable 
Gens point grasping not possible 
etermination 
Finding 
grasping 
opportunities 
  
  
  
  
  
  
Grasp object = 
IAPRS, Vol. 30, Part 5W1, ISPRS Intercommission Workshop "From Pixels to Sequences", Zurich, March 22-24 1995
	        
Waiting...

Note to user

Dear user,

In response to current developments in the web technology used by the Goobi viewer, the software no longer supports your browser.

Please use one of the following browsers to display this page correctly.

Thank you.