View Proposal


Proposer
Ioannis Konstas
Title
NLP - Text Summarisation using Large Language Models
Goal
To build a summarisation system with an LLM
Description
You have most likely already used ChatGPT a few times (or a lot!) Have you ever wondered what it actually takes to build a system based on a Large Language Model (LLM) and evaluate it on a real-world task? In this series of projects (check the rest as well!) we will explore the task of text summarisation. This involves taking a long textual input (it could be a news article, notes, or minutes from an interaction like a meeting, or a dialogue, and converting it into a smaller concise document, or set of bullet points. The important thing is to make sure that the output summary faithfully corresponds to the input while mentioning the most salient/noteworthy points. The idea is to explore several popular techniques for fine-tuning an open-source LLM (e.g., Llama 2) starting from the simpler ones (prompt engineering), all the way up to Parameter Efficient Fine-Tuning (PEFT). We will use standard benchmark datasets and SOTA frameworks to evaluate (and potentially train) our models. We can co-develop the project to emphasise more on the features (conciseness, faithfulness, salience), training, data annotation, or human evaluation.
Resources
Large Language Models, GPU
Background
Machine Learning, NLP (desired: F29AI), software development, F21NL (MSc only), F21CA (MSc only)
Url
Difficulty Level
Challenging
Ethical Approval
None
Number Of Students
1
Supervisor
Ioannis Konstas
Keywords
machine learning, neural networks, large language models, natural language processing (nlp)
Degrees
Bachelor of Science in Computer Science
Master of Science in Artificial Intelligence
Master of Science in Artificial Intelligence with SMI