View Proposal


Proposer
Matthew Aylett
Title
Robot Voice Separation using LSTMs
Goal
Given and audio signal with a mixed 'robot' voice and dialog partner cancel out the robot voice.
Description
Current robots (furhat, haru etc) typically stop listening to their microphone when speaking to avoid interfering with speech recognition results. This means robots can't be interrupted and can't produce back channels (the yeah/okay that shows you are listening). In this project you will extend a small audio corpus built to test robot voice separation, set up the Kaldi ASR system, design and train an neural net solution for altering ASR parameters to remove the effect of the robot voice and maintain recognition accuracy.
Resources
Kaldi, baseline corpora
Background
Aylett, M. P., & Romeo, M. (2023, July). You Don’t Need to Speak, You Need to Listen: Robot Interaction and Human-Like Turn-Taking. In Proceedings of the 5th International Conference on Conversational User Interfaces (pp. 1-5).
Url
Difficulty Level
High
Ethical Approval
None
Number Of Students
2
Supervisor
Matthew Aylett
Keywords
social robotics, speech technology, dialog systems, conversational user interfaces
Degrees
Bachelor of Science in Computer Science
Master of Science in Artificial Intelligence
Master of Science in Data Science
Master of Science in Human Robot Interaction
Master of Science in Robotics