项目作者: facebookresearch

项目描述 :
COVID deterioration prediction based on chest X-ray radiographs via MoCo-trained image representations
高级语言: Python
项目地址: git://github.com/facebookresearch/CovidPrognosis.git
创建时间: 2020-12-10T20:04:03Z
项目社区:https://github.com/facebookresearch/CovidPrognosis

开源协议:MIT License

下载


CovidPrognosis

This repository contains code for reproducing the following paper:

A. Sriram, M. Muckley, K. Sinha, F. Shamout, J. Pineau, K. J. Geras, L. Azour, Y. Aphinyanaphongs, N. Yakubova, W. Moore. COVID-19 Prognosis via Self-Supervised Representation Learning and Multi-Image Prediction. arXiv preprint arXiv:2101.04909 (2020).

We also include models from the MoCo pretraining process for groups interested
in fine-tuning them on their own data. Prior to using this code or pretrained
models please consult the Disclaimer.

Installation

First, follow the
official instructions for
installing PyTorch. Then, navigate to the root CovidPrognosis directory and
run

  1. pip install -e .

After that you should be able to run the examples in cp_examples.

Usage

For pretraining, you’ll need to download the
MIMIC-CXR-JPG or
CheXpert datasets.
Once you’ve downloaded the data, add the path to configs/data.yaml and it
should be used as a default.

The cp_examples directory contains three subdirectories corresponding to the
training stages in the paper:

  • moco_pretrain: Momentum-contrast (MoCo) pretraining (e.g., with MIMIC,
    CheXpert, or both)
  • sip_finetune: Fine-tuning of MoCo models for single-image prediction tasks
    (i.e., single-image adverse event prediction or oxygen requirements prediction)
  • mip_finetune: Fine-tuning of MoCo models for multi-image prediction tasks

Our code is built on top of the
PyTorch Lightning framework.

The examples scripts for MoCo pretraining and SIP fine-tuning are set up for
public X-ray data sets - due to consideration of patient privacy we do not
release COVID data and use the public data as examples.

Pretrained Models

We provide pretrained models that use the
MIMIC-CXR-JPG and
CheXpert
datasets. These datasets are subject to additional terms and conditions as
indicated on their respective websites. For a list of models, see
here. For an example of how to download and train with
the models, please look at the
SIP Fine-tuning example.

The following publication describes MIMIC-CXR:

MIMIC-CXR, a de-identified publicly available database of chest radiographs with free-text reports (A.E.W. Johnson et al., 2019)

And this publication describes CheXpert:

CheXpert: A large chest radiograph dataset with uncertainty labels and expert comparison (J. Irvin et al., 2019)

Disclaimer

This code and accompanying pretrained models are provided with no guarantees
regarding their reliability, accuracy or suitability for any particular
application and should be used for research purposes only. The models and code
are not to be used for public health decisions or responses, or for any
clinical application or as a substitute for medical advice or guidance.

Citation

If you use this code or models in your scientific work, please cite the
following paper:

  1. @misc{sriram2021covid19,
  2. title={COVID-19 Deterioration Prediction via Self-Supervised Representation Learning and Multi-Image Prediction},
  3. author={Anuroop Sriram and Matthew Muckley and Koustuv Sinha and Farah Shamout and Joelle Pineau and Krzysztof J. Geras and Lea Azour and Yindalon Aphinyanaphongs and Nafissa Yakubova and William Moore},
  4. year={2021},
  5. eprint={2101.04909},
  6. archivePrefix={arXiv},
  7. primaryClass={cs.CV}
  8. }

License

The code is MIT licensed, as found in the LICENSE file.