MAIRA-1: A specialised large multimodal model for radiology report generation
- Stephanie Hyland ,
- Shruthi Bannur ,
- Kenza Bouzid ,
- Daniel Coelho de Castro ,
- Mercy Ranjit ,
- Anton Schwaighofer ,
- Fernando Pérez-García ,
- Valentina Salvatelli ,
- Shaury Srivastav ,
- Anja Thieme ,
- Noel Codella ,
- Matthew P Lungren ,
- Maria Teodora Wetscherek ,
- Ozan Oktay ,
- Javier Alvarez-Valle
MSR-TR-2023-47 |
Published by Microsoft
We present a radiology-specific multimodal model for the task for generating radiological reports from chest X-rays (CXRs). Our work builds on the idea that large language model(s) can be equipped with multimodal capabilities through alignment with pre-trained vision encoders. On natural images, this has been shown to allow multimodal models to gain image understanding and description capabilities. Our proposed model (MAIRA-1) leverages a CXR-specific image encoder in conjunction with a fine-tuned large language model based on Vicuna-7B, and text-based data augmentation, to produce reports with state-of-the-art quality. In particular, MAIRA-1 significantly improves on the radiologist-aligned RadCliQ metric and across all lexical metrics considered. Manual review of model outputs demonstrates promising fluency and accuracy of generated reports while uncovering failure modes not captured by existing evaluation practices. More information and resources can be found on the project website: this https URL.