Learning Calibrated Medical Image Segmentation via Multi-rater Agreement Modeling accepted by CVPR 2021.
As depicted in the figure above, in medical image analysis, it is typical to collect multiple annotations, each from a different clinical expert or rater, in the expectation that possible diagnostic errors could be mitigated. Meanwhile, from the computer vision practitioner viewpoint, it has been a common practice to adopt the ground-truth labels obtained via either the majority-vote or simply one annotation from a preferred rater. This process, however, tends to overlook the rich information of agreement or disagreement ingrained in the raw multirater annotations. To address this issue, we propose to explicitly model the multi-rater (dis-)agreement, i.e., MRNet, which effectively improves the calibrated performance for generic medical image segmentation tasks.
- pytorch 1.0.0+
- torchvision
- PIL
- numpy
- tensorboard==1.7.0
- tensorboardX==2.0
Notes for parameters in demo.py
:
- set 'max_iteration' as **Number** # e.g. 2000000, which means the maximum of iteration
- set ‘spshot’ as **Number** # e.g. 20000, which means saving checkpoint every 20000 iterations
- set 'nclass' as **Number** # e.g. 2, which means binary tasks. i.e. model output
- set 'b_size' as **Number** # e.g. 2, whcih means batch size for training
- set 'sshow' as **Number** # e.g. 20, which means showing the training loss every 20 iterations
- set '--phase' as **Train or Test**
- set '--param' as **True or False** # whether load checkpoint or not
- set '--dataset' as **test_name** # set test or val dataset
- set '--snap_num' as **Number** # e.g. 80000, load 80000th checkpoint
- set 'gpu_ids' as **String** # e.g. '0,1', which means running on 0 and 1 GPUs
- You need to set "train_data" and "test_data" path in demo.py
More details in demo.py
- train
python demo.py
# set '--phase' as train
- test
python demo.py
# set '--phase' as test
- Load the log file
cat ./Out/log/_*_.log
- Load the training details
tensorboard --logdir=/YourComputer/model_template/runs/_*_.local/
- RIGA benchmark: you can access to this download link, with fetch code (1627) or Google Drive.
- QUBIQ challenge: the formal web link is here.
@InProceedings{Ji_2021_MRNet,
author = {Ji, Wei and Yu, Shuang and Wu, Junde and Ma, Kai and Bian, Cheng and Bi, Qi and Li, Jingjing and Liu, Hanruo and Cheng, Li and Zheng, Yefeng},
title = {Learning Calibrated Medical Image Segmentation via Multi-Rater Agreement Modeling},
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2021},
pages = {12341-12351}
}
If you have any questions, please contact us ( wji3@ualberta.ca ).