SRG3: Speech-driven Robot Gesture Generation with GAN - ENSTA Paris - École nationale supérieure de techniques avancées Paris Accéder directement au contenu
Communication Dans Un Congrès Année : 2020

SRG3: Speech-driven Robot Gesture Generation with GAN

Résumé

The human gestures occur spontaneously and usually they are aligned with speech, which leads to a natural and expressive interaction. Speech-driven gesture generation is important in order to enable a social robot to exhibit social cues and conduct a successful human-robot interaction. In this paper, the generation process involves mapping acoustic speech representation to the corresponding gestures for a humanoid robot. The paper proposes a new GAN (Generative Adversarial Network) architecture for speech to gesture generation. Instead of the fixed mapping from one speech to one gesture pattern, our end-to-end GAN structure can generate multiple mapped gestures patterns from one speech (with multiple noises) just like humans do. The generated gestures can be applied to social robots with arms. The evaluation result shows the effectiveness of our generative model for speech-driven robot gesture generation.
Fichier principal
Vignette du fichier
ICARCV_2020.pdf (1.8 Mo) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-03047565 , version 1 (08-12-2020)

Identifiants

  • HAL Id : hal-03047565 , version 1

Citer

Chuang Yu, Adriana Tapus. SRG3: Speech-driven Robot Gesture Generation with GAN. 16th International Conference on Control, Automation, Robotics and Vision, Dec 2020, Shenzhen (virtual), China. ⟨hal-03047565⟩
104 Consultations
325 Téléchargements

Partager

Gmail Facebook X LinkedIn More