An integrated system for teaching new visually grounded words to a robot for non-expert users using a mobile device - ENSTA Paris - École nationale supérieure de techniques avancées Paris Accéder directement au contenu
Communication Dans Un Congrès Année : 2009

An integrated system for teaching new visually grounded words to a robot for non-expert users using a mobile device

Résumé

In this paper, we present a system allowing non- expert users to teach new words to their robot. In opposition to most of existing works in this area which focus on the associated visual perception and machine learning challenges, we choose to focus on the HRI challenges with the aim to show that it may improve the learning quality. We argue that by using mediator objects and in particular a handheld device, we can develop a human-robot interface which is not only intuitive and entertaining but will also "help" the user to provide "good" learning examples to the robot and thus will improve the efficiency of the whole learning system. The perceptual and machine learning parts of this system rely on an incremental version of visual bag-of-words. We also propose a system that makes it possible for the robot to incrementally build a model of a novel unknown object by simultaneously modelling and tracking it. We report experiments demonstrating the fast acquisition of robust object models using this approach.
Fichier non déposé

Dates et versions

inria-00420249 , version 1 (28-09-2009)

Identifiants

  • HAL Id : inria-00420249 , version 1

Citer

Pierre-Yves Oudeyer, Pierre Rouanet, David Filliat. An integrated system for teaching new visually grounded words to a robot for non-expert users using a mobile device. IEEE-RAS International Conference on Humanoid Robots, 2009, Tsukuba, Japan. ⟨inria-00420249⟩
131 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More