WinkTalk: A Demonstration of a Multimodal Speech Synthesis Platform Linking Facial Expressions to Expressive Synthetic Voices
File Type:
PDFItem Type:
Conference PaperDate:
2012Access:
openAccessCitation:
Székely, É., Ahmed, Z., Cabral, J.P. & Carson-Berndsen, J., WinkTalk: A Demonstration of a Multimodal Speech Synthesis Platform Linking Facial Expressions to Expressive Synthetic Voices, the Third Workshop on Speech and Language Processing for Assistive Technologies, Montreal, Canada, 7 June, Association for Computational Linguistics, 2012, 5 - 8Download Item:
Abstract:
This paper describes a demonstration of the
WinkTalk system, which is a speech synthesis
platform using expressive synthetic voices.
With the help of a webcamera and facial expression
analysis, the system allows the user
to control the expressive features of the synthetic
speech for a particular utterance with
their facial expressions. Based on a personalised
mapping between three expressive synthetic
voices and the users facial expressions,
the system selects a voice that matches their
face at the moment of sending a message.
TheWinkTalk system is an early research prototype
that aims to demonstrate that facial
expressions can be used as a more intuitive
control over expressive speech synthesis than
manual selection of voice types, thereby contributing
to an improved communication experience
for users of speech generating devices.
Author's Homepage:
http://people.tcd.ie/cabraljOther Titles:
the Third Workshop on Speech and Language Processing for Assistive TechnologiesPublisher:
Association for Computational LinguisticsType of material:
Conference PaperCollections
Availability:
Full text availableSubject (TCD):
Creative Technologies , Digital Engagement , Multimodal System , Speech synthesisMetadata
Show full item recordLicences: