TY - JOUR
T1 - GRTr
T2 - Generative-Retrieval Transformers for Data-Efficient Dialogue Domain Adaptation
AU - Shalyminov, Igor
AU - Sordoni, Alessandro
AU - Atkinson, Adam
AU - Schulz, Hannes
N1 - Funding Information:
Manuscript received October 14, 2020; revised February 5, 2021; accepted April 11, 2021. Date of publication April 21, 2021; date of current version August 6, 2021. This work was supported by the internship at Microsoft Research Montral. The associate editor coordinating the review of this manuscript and approving it for publication was Dr. C. Gunasekara. (Corresponding author: Igor Shalyminov.) Igor Shalyminov is with the Heriot-Watt University, Edinburgh EH14 4AS, U.K. (e-mail: [email protected]).
Publisher Copyright:
© 2014 IEEE.
PY - 2021
Y1 - 2021
N2 - Domain adaptation has recently become a key problem in dialogue systems research. Deep learning, while being the preferred technique for modeling such systems, works best given massive training data. However, in real-world scenarios, such resources are rarely available for new domains, and the ability to train with a few dialogue examples can be considered essential. Pre-training on large data sources and adapting to the target data has become the standard method for few-shot problems within the deep learning framework. In this paper, we present grtr, a hybrid generative-retrieval model based on the large-scale general-purpose language model GPT[2] fine-tuned to the multi-domain metalwoz dataset. In addition to robust and diverse response generation provided by the GPT[2], our model is able to estimate generation confidence, and is equipped with retrieval logic as a fallback for the cases when the estimate is low. grtr is the winning entry at the fast domain adaptation task of DSTC-8 in human evaluation (>4% improvement over the 2nd place system). It also attains superior performance to a series of baselines on automated metrics on metalwoz and multiwoz, a multi-domain dataset of goal-oriented dialogues. In this paper, we also conduct a study of grtr's performance in the setup of limited adaptation data, evaluating the model's overall response prediction performance on metalwoz and goal-oriented performance on multiwoz.
AB - Domain adaptation has recently become a key problem in dialogue systems research. Deep learning, while being the preferred technique for modeling such systems, works best given massive training data. However, in real-world scenarios, such resources are rarely available for new domains, and the ability to train with a few dialogue examples can be considered essential. Pre-training on large data sources and adapting to the target data has become the standard method for few-shot problems within the deep learning framework. In this paper, we present grtr, a hybrid generative-retrieval model based on the large-scale general-purpose language model GPT[2] fine-tuned to the multi-domain metalwoz dataset. In addition to robust and diverse response generation provided by the GPT[2], our model is able to estimate generation confidence, and is equipped with retrieval logic as a fallback for the cases when the estimate is low. grtr is the winning entry at the fast domain adaptation task of DSTC-8 in human evaluation (>4% improvement over the 2nd place system). It also attains superior performance to a series of baselines on automated metrics on metalwoz and multiwoz, a multi-domain dataset of goal-oriented dialogues. In this paper, we also conduct a study of grtr's performance in the setup of limited adaptation data, evaluating the model's overall response prediction performance on metalwoz and goal-oriented performance on multiwoz.
KW - Adaptation models
KW - Context modeling
KW - Data models
KW - Deep learning
KW - dialogue systems
KW - domain adaptation
KW - Gold
KW - natural language processing
KW - neural networks
KW - Predictive models
KW - Task analysis
KW - Training
UR - http://www.scopus.com/inward/record.url?scp=85104642179&partnerID=8YFLogxK
U2 - 10.1109/TASLP.2021.3074779
DO - 10.1109/TASLP.2021.3074779
M3 - Article
AN - SCOPUS:85104642179
SN - 2329-9290
VL - 29
SP - 2484
EP - 2492
JO - IEEE/ACM Transactions on Audio Speech and Language Processing
JF - IEEE/ACM Transactions on Audio Speech and Language Processing
ER -