Synthesising Personality with Neural Speech Synthesis

Shilin Gao, Matthew P. Aylett, David A. Braude, Catherine Lai

Research output: Chapter in Book/Report/Conference proceedingConference contribution

Abstract

Matching the personality of conversational agent to the personality of the user can significantly improve the user experience, with many successful examples in text-based chatbots. It is also important for a voice-based system to be able to alter the personality of the speech as perceived by the users. In this pilot study, fifteen voices were rated using Big Five personality traits. Five content-neutral sentences were chosen for the listening tests. The audio data, together with two rated traits (Extroversion and Agreeableness), were used to train a neural speech synthesiser based on one male and one female voices. The effect of altering the personality trait features was evaluated by a second listening test. Both perceived extroversion and agreeableness in the synthetic voices were affected significantly. The controllable range was limited due to a lack of variance in the source audio data. The perceived personality traits correlated with each other and with the naturalness of the speech. Future work can be making a chatbot speak in a voice with a pre-defined or adaptive personality by using personality synthesis in speech together with text-based personality generation.
Original languageEnglish
Title of host publicationProceedings of the 24th Annual Meeting of the Special Interest Group on Discourse and Dialogue
PublisherAssociation for Computational Linguistics
Pages393-399
Number of pages7
ISBN (Print)9798891760288
DOIs
Publication statusPublished - 2023

Fingerprint

Dive into the research topics of 'Synthesising Personality with Neural Speech Synthesis'. Together they form a unique fingerprint.

Cite this