TY - GEN
T1 - Learning non-cooperative behaviour for dialogue agents
AU - Efstathiou, Ioannis
AU - Lemon, Oliver
PY - 2014
Y1 - 2014
N2 - Non-cooperative dialogue behaviour for artificial agents (e.g. deception and information hiding) has been identified as important in a variety of application areas, including education and health-care, but it has not yet been addressed using modern statistical approaches to dialogue agents. Deception has also been argued to be a requirement for high-order intentionality in AI. We develop and evaluate a statistical dialogue agent using Reinforcement Learning which learns to perform non-cooperative dialogue moves in order to complete its own objectives in a stochastic trading game with imperfect information. We show that, when given the ability to perform both cooperative and non-cooperative dialogue moves, such an agent can learn to bluff and to lie so as to win more games. For example, we show that a non-cooperative dialogue agent learns to win 10.5% more games than a strong rule-based adversary, when compared to an optimised agent which cannot perform non-cooperative moves. This work is the first to show how agents can learn to use dialogue in a non-cooperative way to meet their own goals.
AB - Non-cooperative dialogue behaviour for artificial agents (e.g. deception and information hiding) has been identified as important in a variety of application areas, including education and health-care, but it has not yet been addressed using modern statistical approaches to dialogue agents. Deception has also been argued to be a requirement for high-order intentionality in AI. We develop and evaluate a statistical dialogue agent using Reinforcement Learning which learns to perform non-cooperative dialogue moves in order to complete its own objectives in a stochastic trading game with imperfect information. We show that, when given the ability to perform both cooperative and non-cooperative dialogue moves, such an agent can learn to bluff and to lie so as to win more games. For example, we show that a non-cooperative dialogue agent learns to win 10.5% more games than a strong rule-based adversary, when compared to an optimised agent which cannot perform non-cooperative moves. This work is the first to show how agents can learn to use dialogue in a non-cooperative way to meet their own goals.
U2 - 10.3233/978-1-61499-419-0-999
DO - 10.3233/978-1-61499-419-0-999
M3 - Conference contribution
AN - SCOPUS:84923171835
SN - 9781614994183
VL - 263
T3 - Frontiers in Artificial Intelligence and Applications
SP - 999
EP - 1000
BT - Frontiers in Artificial Intelligence and Applications
PB - IOS Press
T2 - 21st European Conference on Artificial Intelligence
Y2 - 18 August 2014 through 22 August 2014
ER -