Using Multi-view Recognition and Meta-data Annotation to Guide a Robot's Attention

Alexander Thomas, Vittorio Ferrari, Bastian Liebe, Tinne Tuytelaars, Luc van Gool

Research output: Contribution to journalArticlepeer-review

Abstract / Description of output

In the transition from industrial to service robotics, robots will have to deal with increasingly unpredictable and variable environments. We present a system that is able to recognize objects of a certain class in an image and to identify their parts for potential interactions. The method can recognize objects from arbitrary viewpoints and generalizes to instances that have never been observed during training, even if they are partially occluded and appear against cluttered backgrounds. Our approach builds on the implicit shape model of Leibe et al. We extend it to couple recognition to the provision of meta-data useful for a task and to the case of multiple viewpoints by integrating it with the dense multi-view correspondence finder of Ferrari et al. Meta-data can be part labels but also depth estimates, information on material types, or any other pixelwise annotation. We present experimental results on wheelchairs, cars, and motorbikes.
Original languageEnglish
Pages (from-to)976-988
Number of pages13
JournalInternational Journal of Robotics Research
Issue number8
Publication statusPublished - Aug 2009


Dive into the research topics of 'Using Multi-view Recognition and Meta-data Annotation to Guide a Robot's Attention'. Together they form a unique fingerprint.

Cite this