ACTIVE LEARNING FOR VISION-BASED ROBOT GRASPING
by Marcos Salganicoff, Lyle H. Ungar, and Ruzena Bajcsy
Reliable vision-based grasping has proved elusive outside of
controlled environments. One approach towards building more flexible
and domain-independent robot grasping systems is to employ learning to
adapt the robot's perceptual and motor system to the task. However,
one pitfall in robot perceptual and motor learning is that the cost of
gathering the learning set may be unacceptably high. Active learning
algorithms address this shortcoming by intelligently selecting actions
so as to decrease the number of examples necessary to achieve good
performance and also avoid separate training and execution phases,
leading to higher autonomy. We describe the IE-ID3 algorithm, which
extends the Interval Estimation (IE) active learning approach from
discrete to real-valued learning domains by combining IE with a
classification tree learning algorithm (ID-3). We present a robot
system which rapidly learns to select the grasp approach directions
using IE-ID3 given simplified superquadric shape approximations of
objects. Initial results on a small set of objects show that a robot
with a laser scanner system can rapidly learn to pick up new objects,
and simulation studies show the superiority of the active learning
approach for a simulated grasping task using larger sets of objects.
Extensions of the approach and future areas of research incorporating
more sophisticated perceptual and action representation are discussed.