To effectively unleash the power of labeled and unlabeled data, we propose to combine self- and semi-supervised learning for omni-supervised pre-training.
Model | Params | Checkpoint |
---|---|---|
VoComni_nnunet | 31M | Download |
VoComni_B | 72M | Download |
VoComni_L | 290M | Download |
VoComni_H | 1.2B | Download |
Please refer to Acknowledgment. Download our PreCT-160K and VoComni for pre-training.
The path of PreCT-160K should be organized as:
# or you can modify it in 'utils/data_utils*.py'
├── data
├── BTCV
├── TCIAcovid19
├── Luna16-jx
├── ...
├── VoComni
└── cache
WARNING:
- It requires about 60 TB space to store and cache the datasets.
cd Self-supervised
source activate YOUR-CONDA-ENVIRONMENT
# single GPU, if you don't have enough gpu resource
sh single_train
# multi-gpu
sh dist_B.sh
sh dist_L.sh
sh dist_H.sh
NOTE THAT we are not the authors of these datasets. Although all these datasets are publicly available for academic research, you need to cite the original works as shown in our paper. For certain datasets (e.g., WORD) that necessitate approval from the authors, you need to download it from the original link.
If you find this repo useful for your research, please consider citing the paper as follows:
@article{wu2024large,
title={Large-Scale 3D Medical Image Pre-training with Geometric Context Priors},
author={Wu, Linshan and Zhuang, Jiaxin and Chen, Hao},
journal={arXiv preprint arXiv:2410.09890},
year={2024}
}
@InProceedings{voco-v1,
author = {Wu, Linshan and Zhuang, Jiaxin and Chen, Hao},
title = {VoCo: A Simple-yet-Effective Volume Contrastive Learning Framework for 3D Medical Image Analysis},
booktitle = {CVPR},
month = {June},
year = {2024},
pages = {22873-22882}
}