A Multimodal Human-Robot Interaction Dataset - ENSTA Paris - École nationale supérieure de techniques avancées Paris Accéder directement au contenu
Poster De Conférence Année : 2016

A Multimodal Human-Robot Interaction Dataset

Résumé

This works presents a multimodal dataset for Human-Robot Interactive Learning. 1 The dataset contains synchronized recordings of several human users, from a stereo 2 microphone and three cameras mounted on the robot. The focus of the dataset is 3 incremental object learning, oriented to human-robot assistance and interaction. To 4 learn new object models from interactions with a human user, the robot needs to 5 be able to perform multiple tasks: (a) recognize the type of interaction (pointing, 6 showing or speaking), (b) segment regions of interest from acquired data (hands and 7 objects), and (c) learn and recognize object models. We illustrate the advantages 8 of multimodal data over camera-only datasets by presenting an approach that 9 recognizes the user interaction by combining simple image and language features.
Fichier principal
Vignette du fichier
multimodal-dataset-nips(1).pdf (6.23 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)
Loading...

Dates et versions

hal-01402479 , version 1 (07-12-2016)

Identifiants

  • HAL Id : hal-01402479 , version 1

Citer

Pablo Azagra, Yoan Mollard, Florian Golemo, Ana Cristina Murillo, Manuel Lopes, et al.. A Multimodal Human-Robot Interaction Dataset. NIPS 2016, workshop Future of Interactive Learning Machines, Dec 2016, Barcelona, Spain. ⟨hal-01402479⟩
646 Consultations
769 Téléchargements

Partager

Gmail Facebook X LinkedIn More