Skip to content

Claire1217/AGPT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

15 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Anatomical Grounding Pre-training for Medical Phrase Grounding

Abstract

Medical Phrase Grounding (MPG) maps radiological findings to specific regions in medical images, aiding in the understanding of radiology reports. Existing generaldomain models struggle in zero-shot and fine-tuning settings with MPG due to domain-specific challenges and limited annotated data. We propose anatomical grounding as an in-domain pre-training task that aligns anatomical terms with corresponding regions in medical images, leveraging large-scale datasets such as Chest ImaGenome. Our emperical evaluation on MS-CXR demonstrates that anatomical grounding pre-training significantly improves performance in both a zero-shot and fine-tuning setting, outperforming selfsupervised models like GLoRIA and BioViL. Our fine-tuned model achieved state-of-the-art performance on MS-CXR with an mIoU of 61.2, demonstrating the effectiveness of anatomical grounding pre-training for MPG.

Description

This repository provides the code necessary to load the MDETR and TransVG models, which are pre-trained on the ChestImagenome anatomy grounding dataset and subsequently fine-tuned on the MS-CXR medical phrase grounding dataset. The demo folder includes four example Chest X-ray images from the MS-CXR dataset. The demo.py script supplies descriptive phrases along with their corresponding bounding box annotations for each image. Additionally, the repository offers visualizations that display the model's predictions, highlighting the annotated bounding boxes on the sample Chest X-ray images.

Model weights

The fine-tuned model weights for TransVG-AGPT and MDETR-AGPT on the MS-CXR dataset are available for download. Google Drive Link:https://drive.google.com/drive/folders/1lFaqAlG6AJExCjkW3Pbb24CX-LWSstrK?usp=sharing. Save them in: agpt/model_weight/transvg.pth agpt/model_weight/mdetr.pth

Demonstration

We strongly recommend using Python 3.10 to run this demonstration, as other versions may cause conflicts between packages. After Python 3.10 is installed on your system, you can then install the required packages using:

pip install -r requirements.txt

To run the demonstration, execute:

python demo.py

This script loads the trained MDETR and TransVG models to perform predictions on four example images located in the demo/ directory, with their corresponding phrases and bounding boxes defined within demo.py. The results will be visualized and saved in demo/outputs/mdetr and demo/outputs/transvg directories. In the visualizations, ground truth bounding boxes are displayed in red and predicted bounding boxes in blue.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages