View Proposal


Proposer
Matthew Aylett
Title
Graphical character for conversational interaction
Goal
Implement a graphical character that is pollable, streamable and interruptible
Description
In human dialog, participants are able to interrupt each other at any point. Once a dialog participant has been interrupted they may cede the floor (let another person speak) or alter what they are saying to show they are actively listening. In this project a graphical character will be implemented using Unity and Python that allows some limited behaviour (e.g. head nods, eye brow raise), that is integrated with a speech synthesis system (CereProc) allowing basic lip syncing and output speech. An API will be designed and built over http that will allow a client Python program to stream content (e.g. text for the character to speak and instructions for behaviour), poll the character (establish within a narrow time window what the character has said or done from the streamed content), and interruptible (will allow the system to stop the character gracefully or to stop the character and begin producing new content).
Resources
CereProc TTS system
Background
Aylett, M. P., & Romeo, M. (2023, July). You Don’t Need to Speak, You Need to Listen: Robot Interaction and Human-Like Turn-Taking. In Proceedings of the 5th International Conference on Conversational User Interfaces (pp. 1-5).
Url
Difficulty Level
Challenging
Ethical Approval
Full
Number Of Students
2
Supervisor
Matthew Aylett
Keywords
face animation, python
Degrees
Bachelor of Science in Computer Science
Master of Design in Games Design and Development
Master of Science in Artificial Intelligence
Master of Science in Human Robot Interaction
Master of Science in Robotics