Dual Pseudo-Labels Interactive Self-Training for Semi-Supervised Visible-Infrared Person Re-Identification
This is an official implementation of "Dual Pseudo-Labels Interactive Self-Training for Semi-Supervised Visible-Infrared Person Re-Identification", accepted by ICCV 2023. Paper Link
- We propose a dual pseudo-label interactive self-training framework for semi-supervised visible-infrared person Re-ID, which leverages the intro- and inter-modality characteristics to obtain hybrid pseudo-labels.
- We introduce three modules: noise label penalty (NLP), noise correspondence calibration (NCC), and unreliable anchor learning (UAL). These modules help to penalize noise labels, calibrate noisy correspondences, and exploit hard-to-discriminate features.
- We provide comprehensive evaluations under these two semi-supervised VI-ReID. Extensive experiments on two popular VI-ReID benchmarks demonstrate that our DPIS achieves impressive performance.
1. sh run_train_sysu.sh for SYSU-MM01
2. sh run_train_regdb.sh for RegDB
1. sh run_test_sysu.sh for SYSU-MM01
2. sh run_test_regdb.sh for RegDB
If our work is helpful for your research, please consider citing:
@inproceedings{shi2023dual,
title={Dual Pseudo-Labels Interactive Self-Training for Semi-Supervised Visible-Infrared Person Re-Identification},
author={Shi, Jiangming and Zhang, Yachao and Yin, Xiangbo and Xie, Yuan and Zhang, Zhizhong and Fan, Jianping and Shi, Zhongchao and Qu, Yanyun},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={11218--11228},
year={2023}
}
jiangming.shi@outlook.com; S_yinxb@163.com.
The code is implemented based on OTLA.