View Proposal
-
Proposer
-
John See
-
Title
-
Talk to the Museum: Multimodal Retrieval Augmented Generation for Museums
-
Goal
-
A chat application that allows museum visitors to query for more information about things in a museum or heritage site using text and pictures
-
Description
- Exploring new interactions between visitors and the museum may be vital to increase the appeal of museums (and heritage sites) for the new generation. Traditionally, museums tend to depend a lot on manually disseminated information such as site experts and tour guides, as well as using specific types of technologies like sensor-based audio/visual guides and on-site augmented reality to make things interesting. Recent advances in AI-based chatbots now present new possibilities to museums. Wouldn't it be fascinating if visitors could ask questions to the chatbot about the historical facts of a specific artifact in the museum? Can visitors ask the chatbot about a certain picture of a specific pattern on the exhibit?
This project should explore the use of retrieval augmented generation (RAG), which combines the strengths of traditional information retrieval systems with the capabilities of generative large language models (LLMs), build a demonstrable prototype, and have it evaluated by early users.
- Resources
-
Free cloud compute (Colab) should suffice. Application can be hosted locally or publicly (HuggingFace, etc.)
-
Background
-
Reasonable level of competency in programming, especially Python; Some familiarity with LLMs and API integration in apps would be an added advantage;
-
Url
-
-
Difficulty Level
-
Moderate
-
Ethical Approval
-
InterfaceOnly
-
Number Of Students
-
1
-
Supervisor
-
John See
-
Keywords
-
multimodal llms, retrieval augmented generation, generative ai, multimedia for heritage
-
Degrees
-
Bachelor of Science in Computing Science