View Proposal
-
Proposer
-
Gavin Abercrombie
-
Title
-
Safety and Bias in Dialogue Systems and Large Language Models
-
Goal
-
Evaluate and develop datasets, methods, and systems for the detection and mitigation of safety issues and biases in Dialogue Systems and LLMs.
-
Description
- Automated dialogue systems are becoming ubiquitous in our homes, on our smart devices, and on the internet, and, with recent advances in Large Language Models, the quality of chatbots and voice assistants is rapidly improving — to the extent that they can sometimes be mistaken for humans [1].
But as the quality of end-to-end dialogue systems improves, so does their capacity to learn unsafe behaviours from data on which they are trained [2]. They also run the risk of responding inappropriately to unsafe or toxic user input [3].
Potentially undesirable behaviours include offensive outputs (abuse, hate speech etc.), as well as the failure to detect and mitigate such language in user inputs, and generation of inappropriate responses in safety-critical situations, such as offering medical or legal advice, or responding to/generating sensitive content.
In this project, we will examine one or more of the following aspects of safety for conversational AI:
- Evaluation and/or detection of unsafe user input and/or system outputs (abusive language, safety critical topics, sensitive content etc.).
- Evaluation of societal biases in the outputs of conversational systems and Large Language Models (LLMs)
- Mitigation of unsafe content e.g. evaluation of system response strategies, generation of appropriate responses.
- Resources
-
[1] Abercrombie, G., Cercas Curry, A., Pandya, M., & Rieser, V. (2021). Alexa, Google, Siri: What are Your Pronouns? Gender and Anthropomorphism in the Design and Perception of Conversational Assistants. Proceedings of the 3rd Workshop on Gender Bias in Natural Language Processing.
[2] Dinan, E., Gavin Abercrombie, A. Bergman, Shannon Spruit, Dirk Hovy, Y-Lan Boureau, and Verena Rieser. 2022. Safety Kit: First Aid for Measuring Safety in Open-Domain Conversational Systems. ACL.
-
Background
-
-
Url
-
-
Difficulty Level
-
Moderate
-
Ethical Approval
-
Full
-
Number Of Students
-
0
-
Supervisor
-
Gavin Abercrombie
-
Keywords
-
safety, bias, nlp, dialogue systems, large language models
-
Degrees
-
Master of Science in Artificial Intelligence
Master of Science in Artificial Intelligence with SMI
Master of Science in Data Science