A Multimodal Dataset for Interactive and Incremental Learning of Object Models - ENSTA Paris - École nationale supérieure de techniques avancées Paris Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2016

A Multimodal Dataset for Interactive and Incremental Learning of Object Models

Résumé

This work presents an incremental object learning framework oriented to human-robot assistance and interaction. To learn new object models from interactions with a human user, the robot needs to be able to perform multiple recognition tasks: (a) recognize the type of interaction, (b) segment regions of interest from acquired data, and (c) learn and recognize object models. The contributions on this work are focused on the recognition modules of this human-robot interactive framework. First, we illustrate the advantages of multimodal data over camera-only datasets. We present an approach that recognizes the user interaction by combining simple image and language features. Second, we propose an incremental approach to learn visual object models, which is shown to achieve comparable performance to a typical offline-trained system. We utilize two public datasets, one of them presented and released in this work. This dataset contains synchronized recordings from user speech and three cameras mounted on a robot, which captured the user teaching object names to the robot.
Fichier principal
Vignette du fichier
multimodal-dataset-interactive.pdf (8.83 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01402493 , version 1 (24-11-2016)

Identifiants

  • HAL Id : hal-01402493 , version 1

Citer

Pablo Azagra, Yoan Mollard, Florian Golemo, Ana C Murillo, Manuel Lopes, et al.. A Multimodal Dataset for Interactive and Incremental Learning of Object Models. 2016. ⟨hal-01402493⟩
437 Consultations
403 Téléchargements

Partager

Gmail Facebook X LinkedIn More