Abstract
Multimodal search-based dialogue is a challenging new task: It extends visually grounded question answering systems into multi-turn conversations with access to an external database. We address this new challenge by learning a neural response generation system from the recently released Multimodal Dialogue (MMD) dataset (Saha et al., 2017). We introduce a knowledge-grounded multimodal conversational model where an encoded knowledge base (KB) representation is appended to the decoder input. Our model substantially outperforms strong baselines in terms of text-based similarity measures (over 9 BLEU points, 3 of which are solely due to the use of additional information from the KB).
Original language | English |
---|---|
Title of host publication | Proceedings of the 2018 EMNLP Workshop SCAI |
Subtitle of host publication | The 2nd International Workshop on Search-Oriented Conversational AI |
Publisher | Association for Computational Linguistics |
Pages | 59-66 |
Number of pages | 8 |
ISBN (Electronic) | 9781948087759 |
Publication status | Published - 31 Oct 2018 |
Event | 2nd International Workshop on Search-Oriented Conversational AI - Brussels, Belgium Duration: 31 Oct 2018 → 31 Oct 2018 |
Workshop
Workshop | 2nd International Workshop on Search-Oriented Conversational AI |
---|---|
Country/Territory | Belgium |
City | Brussels |
Period | 31/10/18 → 31/10/18 |