ConvAbuse: Data, Analysis, and Benchmarks for Nuanced Abuse Detection in Conversational AI

Amanda Cercas Curry*, Gavin Abercrombie, Verena Rieser

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

We present the first English corpus study on abusive language towards three conversational AI systems gathered ‘in the wild’: an open-domain social bot, a rule-based chatbot, and a task-based system. To account for the complexity of the task, we take a more ‘nuanced’ approach where our ConvAI dataset reflects fine-grained notions of abuse, as well as views from multiple expert annotators. We find that the distribution of abuse is vastly different compared to other commonly used datasets, with more sexually tinted aggression towards the virtual persona of these systems. Finally, we report results from bench-marking existing models against this data. Unsurprisingly, we find that there is substantial room for improvement with F1 scores below 90%.
Original languageEnglish
Title of host publicationProceedings of the 2021 Conference on Empirical Methods in Natural Language Processing
PublisherAssociation for Computational Linguistics
Pages7388–7403
Number of pages16
ISBN (Print) 9781955917094
DOIs
Publication statusPublished - Nov 2021
Event2021 Conference on Empirical Methods in Natural Language Processing - Virtual, Punta Cana, Dominican Republic
Duration: 7 Nov 202111 Nov 2021

Conference

Conference2021 Conference on Empirical Methods in Natural Language Processing
Abbreviated titleEMNLP 2021
Country/TerritoryDominican Republic
CityVirtual, Punta Cana
Period7/11/2111/11/21

Fingerprint

Dive into the research topics of 'ConvAbuse: Data, Analysis, and Benchmarks for Nuanced Abuse Detection in Conversational AI'. Together they form a unique fingerprint.

Cite this