This is the official implementation of "Geometry Uncertainty Projection Network for Monocular 3D Object Detection".
An updated version of the GUPNet, called GUPNet++, has been released at GUPNET++. We hope this new version can contribute to your research. Regards.
If you find our work useful in your research, please consider citing:
@article{lu2021geometry,
title={Geometry Uncertainty Projection Network for Monocular 3D Object Detection},
author={Lu, Yan and Ma, Xinzhu and Yang, Lei and Zhang, Tianzhu and Liu, Yating and Chu, Qi and Yan, Junjie and Ouyang, Wanli},
journal={arXiv preprint arXiv:2107.13774},year={2021}}
Download the KITTI dataset from KITTI website, including left color images, camera calibration matrices and training labels.
Clone this project and then go to the code directory:
git clone https://github.com/SuperMHP/GUPNet.git
cd code
We train the model on the following environments:
Python 3.6
Pytorch 1.1
Cuda 9.0
You can build the environment easily by installing the requirements:
conda env create -f requirements.yml
conda activate gupnet
Train the model:
CUDA_VISIBLE_DEVICES=0,1,2 python tools/train_val.py
After training the model will directly feedback the detection files for evaluation (If so, you can skip this setep). But if you want to test a given checkpoint, you need to modify the "resume" of the "tester" in the code/experiments/config.yaml and then run:
python tools/train_val.py -e
After that, please use the kitti evaluation devkit (deails can be refered to FrustumPointNet) to evaluate:
g++ evaluate_object_3d_offline_apXX.cpp -o evaluate_object_3d_offline_ap
../../tools/kitti_eval/evaluate_object_3d_offline_apXX KITTI_LABEL_DIR ./output
We also provide the trained checkpoint which achieved the best multi-category performance on the validation set. It can be downloaded at here. This checkpoint performance is as follow:
Models | Car@IoU=0.7 | Pedestrian@IoU=0.5 | Cyclist@IoU=0.5 | ||||||
Easy | Mod | Hard | Easy | Mod | Hard | Easy | Mod | Hard | |
original paper | 22.76% | 16.46% | 13.72% | - | - | - | - | - | - |
released chpt | 23.19% | 16.23% | 13.57% | 11.29% | 7.05% | 6.36% | 9.49% | 5.01% | 4.14% |
Modify the train set to the trainval set (You can modify it in the code/libs/helpers/dataloader_helper.py), and then modify the input of the evaluation function to the test set (code/tools/train_val.py).
Compressed the output file to a zip file (Please note that this zip file do NOT include any root directory):
cd outputs/data
zip -r submission.zip .
submit this file to the KITTI page (You need to register an account.)
We also give our trained checkpoint on the trainval dataset. You can download it from here. This checkpoint performance is as follow (KITTI page):
Models | Car@IoU=0.7 | Pedestrian@IoU=0.5 | Cyclist@IoU=0.5 | ||||||
Easy | Mod | Hard | Easy | Mod | Hard | Easy | Mod | Hard | |
original paper | 20.11% | 14.20% | 11.77% | 14.72% | 9.53% | 7.87% | 4.18% | 2.65% | 2.09% |
released chpt | 22.26% | 15.02% | 13.12% | 14.95% | 9.76% | 8.41% | 5.58% | 3.21% | 2.66% |
-
The releases code is originally set to train on multi-category here. If you would like to train on the single category (Car), please modify the code/experiments/config.yaml. Single-category training can lead to higher performance on the Car.
-
This implementation includes some tricks that do not describe in the paper. Please feel free to ask me in the issue. And I will also update the principle of them in the supplementary materials
-
The overall code cannot completely remove randomness because we use some functions which do not have reproduced implementation (e.g. ROI align). So the performance may have a certain degree of jitter, which is normal for this project.
If you have any question about this project, please feel free to contact yan.lu1@sydney.edu.au.