Complete but Unofficial PyTorch Implementation of Complex-YOLO: Real-time 3D Object Detection on Point Clouds with YoloV3
$ git clone https://github.com/ghimiredhikura/Complex-YOLOv3
$ cd Complex-YOLO-V3/
$ sudo pip install -r requirements.txt
Download pretrained weights [yolov3, tiny-yolov3]
$ cd checkpoints/
$ python download_weights.py
1. $ python test_detection.py --split=sample --folder=sampledata
2. $ python test_both_side_detection.py --split=sample --folder=sampledata
Demo Video [Click to Play]
You can see sampledata
folder in data/KITTI/dataset
directory which can be used for testing this project without downloading KITTI dataset. However, if you want to train the model by yourself and check the mAP in validation set just follow the steps below.
Download the 3D KITTI detection dataset
- Camera calibration matrices of object data set (16 MB)
- Training labels of object data set (5 MB)
- Velodyne point clouds (29 GB)
- Left color images of object data set (12 GB)
Now you have to manage dataset directory structure. Place your dataset into data
folder. Please make sure that you have the dataset directory structure as follows.
└── data/KITTI/object
├── training <-- 7481 train data
| ├── image_2 <-- for visualization
| ├── calib
| ├── label_2
| ├── velodyne
└── testing <-- 7580 test data
├── image_2 <-- for visualization
├── calib
├── velodyne
The train/valid
split of training dataset as well as sample
and test
dataset ids are in data/KITTI/ImageSets
directory. From training set of 7481 images, 6000 images are used for training and remaining 1481 images are used for validation. The mAP results reported in this project are evaluated into this valid set with custom mAP evaluation script with 0.5 iou for each object class.
$ python check_dataset.py
$ train.py [-h] [--epochs EPOCHS] [--batch_size BATCH_SIZE]
[--gradient_accumulations GRADIENT_ACCUMULATIONS]
[--model_def MODEL_DEF]
[--pretrained_weights PRETRAINED_WEIGHTS]
[--n_cpu N_CPU] [--img_size IMG_SIZE]
[--evaluation_interval EVALUATION_INTERVAL]
[--multiscale_training MULTISCALE_TRAINING]
--Training log example--
---- [Epoch 0/300, Batch 250/1441] ----
+------------+--------------+--------------+--------------+
| Metrics | YOLO Layer 0 | YOLO Layer 1 | YOLO Layer 2 |
+------------+--------------+--------------+--------------+
| grid_size | 17 | 34 | 68 |
| loss | 6.952686 | 5.046788 | 4.256296 |
| x | 0.054503 | 0.047048 | 0.060234 |
| y | 0.110871 | 0.059848 | 0.081368 |
| w | 0.101059 | 0.056696 | 0.022349 |
| h | 0.294365 | 0.230845 | 0.076873 |
| im | 0.215230 | 0.218564 | 0.184226 |
| re | 1.049812 | 0.883522 | 0.783887 |
| conf | 4.682138 | 3.265709 | 2.941420 |
| cls | 0.444707 | 0.284557 | 0.105938 |
| cls_acc | 67.74% | 83.87% | 96.77% |
| recall50 | 0.000000 | 0.129032 | 0.322581 |
| recall75 | 0.000000 | 0.032258 | 0.032258 |
| precision | 0.000000 | 0.285714 | 0.133333 |
| conf_obj | 0.058708 | 0.248192 | 0.347815 |
| conf_noobj | 0.014188 | 0.007680 | 0.010709 |
+------------+--------------+--------------+--------------+
Total loss 16.255769729614258
---- ETA 0:18:27.490254
$ python test_detection.py
$ python test_both_side_detection.py
$ python eval_mAP.py
mAP (min. 50 IoU)
Model/Class | Car | Pedestrian | Cyclist | Average |
---|---|---|---|---|
Complex-YOLO-v3 | 97.89 | 82.71 | 90.12 | 90.24 |
Complex-Tiny-YOLO-v3 | 95.91 | 49.29 | 78.75 | 74.65 |
- Complex-YOLO: https://arxiv.org/pdf/1803.06199.pdf
YoloV3 Implementation is borrowed from:
Point Cloud Preprocessing is based on: