Abstract
Non-cooperative dialogue behaviour has been identified as important in a variety of application areas, including education, military operations, video games and healthcare. However, it has not been addressed using statistical approaches to dialogue management, which have always been trained for co-operative dialogue. We develop and evaluate a statistical dialogue agent which learns to perform non-cooperative dialogue moves in order to complete its own objectives in a stochastic trading game. We show that, when given the ability to perform both cooperative and non-cooperative dialogue moves, such an agent can learn to bluff and to lie so as to win games more often -- against a variety of adversaries, and under various conditions such as risking penalties for being caught in deception. For example, we show that a non-cooperative dialogue agent can learn to win an additional 15.47% of games against a strong rule-based adversary, when compared to an optimised agent which cannot perform non-cooperative moves. This work is the first to show how an agent can learn to use non-cooperative dialogue to effectively meet its own goals.
Original language | English |
---|---|
Title of host publication | Proceedings of the 15th Annual Meeting of the Special Interest Group on Discourse and Dialogue |
Publisher | Association for Computational Linguistics |
Pages | 60-68 |
Number of pages | 9 |
ISBN (Print) | 978-1-941643-21-1 |
Publication status | Published - 18 Jun 2014 |
Event | 15th Annual Meeting of the Special Interest Group on Discourse and Dialogue 2014 - Philadelphia, PA, United States Duration: 18 Jun 2014 → 20 Jun 2014 |
Conference
Conference | 15th Annual Meeting of the Special Interest Group on Discourse and Dialogue 2014 |
---|---|
Abbreviated title | SIGDIAL 2014 |
Country/Territory | United States |
City | Philadelphia, PA |
Period | 18/06/14 → 20/06/14 |