View Proposal


Proposer
Heba Elshimy
Title
Zero-Shot Medical Image Classification via Vision-Language Prompt Engineering
Goal
This aims to explore if we can leverage the knowledge embedded in large-scale pre-trained models (VLMs) to achieve clinically useful performance without the costly annotation and training typically required for medical imaging tasks
Description
This project investigates the capability of Vision-Language Models (VLMs) to perform zero-shot and few-shot medical image classification through systematic prompt engineering, without task-specific fine-tuning. As foundation models increasingly demonstrate cross-domain generalisation, understanding their boundaries in specialised medical contexts becomes essential.
Resources
Datasets: CheXpert (https://stanfordmlgroup.github.io/competitions/chexpert/), ISIC Archive (https://www.isic-archive.com/)
Background
Knowledge with huggingface and open source foundational or large-scale models
Url
Difficulty Level
High
Ethical Approval
None
Number Of Students
1
Supervisor
Heba Elshimy
Keywords
vlms, ai for healthcare
Degrees
Master of Science in Artificial Intelligence
Master of Science in Computer Science for Cyber Security
Master of Science in Data Science
Master of Science in Network Security
MSc Applied Cyber Security