Skip to content

cheliu-computation/M-FLAG-MICCAI2023

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

M-FLAG-MICCAI2023

M-FLAG: Medical Vision-Language Pre-training with Frozen Language Models and Latent Space Geometry Optimization, MICCAI 2023.

Installation

To clone this repository:

git clone https://github.com/cheliu-computation/M-FLAG-MICCAI2023.git

To install Python dependencies:

pip install -r requirements.txt

All experiments are implemented on A100 GPU.

Pre-train Dataset downloading

Datasets we used are as follows:

  • MIMIC-CXR: We downloaded the MIMIC-CXR-JPG dataset as the radiographs. Paired medical reports can be downloaded in MIMIC-CXR.

Preprocessing

  • First we follow MGCA preprocessing to extract a master csv includes all CXR scans associated with report. You can find in Preprocessing.
  • Then, run 'ext_data.py' to extract all scans and save as a npy file. It will accelerate the pre-training stage.

Pre-training

We pre-trained MGCA on MIMIC-CXR using this command:


cd M-FLAG-MICCAI2023/pretrain
torchrun --nnodes=1 --nproc_per_node=2 main.py

Finetune on downstream tasks

We evlauate the performance of M-FLAG on three downstream tasks: classification, object detection and semantic segmentation.

For classification task, we follow CheXclusion, please follow their offical code to extract data and implement classification tasks.

For semantic segmentation and object detection, we follow MGCA offical configuration and code. The dataset can be found in MGCA repository.

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages