A Learning-based Co-Speech Gesture Generation System for Social Robots

Xiangqi Li, Christian Dondrup

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Co-speech gestures enhance both human-human and human-robot interactions. This paper examines the efficacy of a data-driven approach for generating synchronised co-speech gestures in three social robots to improve social interactions. Building on a sequence-to-sequence model, which maps speech to gestures [21], this work uses the Talking With Hands 16.2M dataset [11] to generate natural gestures for face-to-face conversations. Additionally, we address synchronisation issues identified in the original study. The model’s generality is tested on three robots—NAO, Pepper, and ARI. Objective and subjective evaluations, confirm that a data-driven approach effectively generates synchronised co-speech gestures.
Original languageEnglish
Title of host publicationHAI '24: Proceedings of the 12th International Conference on Human-Agent Interaction
PublisherAssociation for Computing Machinery
Pages453-455
Number of pages3
ISBN (Print)9798400711787
DOIs
Publication statusPublished - 24 Nov 2024
Event12th International Conference on Human-Agent Interaction 2024
- Swansea University, Swansea, United Kingdom
Duration: 24 Nov 202427 Nov 2024
https://hai-conference.net/hai2024/

Conference

Conference12th International Conference on Human-Agent Interaction 2024
Country/TerritoryUnited Kingdom
CitySwansea
Period24/11/2427/11/24
Internet address

Fingerprint

Dive into the research topics of 'A Learning-based Co-Speech Gesture Generation System for Social Robots'. Together they form a unique fingerprint.

Cite this