This is the official repo of paper Learning RoI Transformer for Detecting Oriented Objects in Aerial Images
This code is based on deformable convolution network
We refactored the code and retrained the model. There are slight differences in the final accuracy.
mmdetection version is finished, it is more faster and accurate, we recommend you to use the new version.
-
MXNet from the offical repository.
-
Python 2.7. We recommend using Anaconda2 as it already includes many common packages. We do not support Python 3 yet, if you want to use Python 3 you need to modify the code to make it work.
-
Python packages might missing: cython, opencv-python >= 3.2.0, easydict. If
pip
is set up on your system, those packages should be able to be fetched and installed by runningpip install -r requirements.txt
-
For Windows users, Visual Studio 2015 is needed to compile cython module.
- Clone the RoI Transformer repository, and we'll call the directory that you cloned RoI Transformer as ${RoI_ROOT}
git clone https://github.com/dingjiansw101/RoITransformer_DOTA.git
-
For Windows users, run
cmd .\init.bat
. For Linux user, runsh ./init.sh
. The scripts will build cython module automatically and create some folders. -
Install MXNet:
Note: The MXNet's Custom Op cannot execute parallelly using multi-gpus after this PR. We strongly suggest the user rollback to version MXNet@(commit 998378a) for training (following Section 3.2 - 3.5).
Build from source (Since there are custom c++ operators, We need to complie the MXNet from source.)
3.1 Clone MXNet and checkout to MXNet@(commit 998378a) by
git clone --recursive https://github.com/dmlc/mxnet.git git checkout 998378a git submodule update # if it's the first time to checkout, just use: git submodule update --init --recursive
3.2 Copy the c++ operators to MXNet source
cp ${RoI_ROOT}/fpn/operator_cxx/* ${MXNET_ROOT}/src/operator/contrib
3.3 Compile MXNet
cd ${MXNET_ROOT} make -j $(nproc) USE_OPENCV=1 USE_BLAS=openblas USE_CUDA=1 USE_CUDA_PATH=/usr/local/cuda USE_CUDNN=1
3.4 Install the MXNet Python binding by
Note: If you will actively switch between different versions of MXNet, please follow 3.5 instead of 3.4
cd python sudo python setup.py install
3.5 For advanced users, you may put your Python packge into
./external/mxnet/$(YOUR_MXNET_PACKAGE)
, and modifyMXNET_VERSION
in./experiments/rfcn/cfgs/*.yaml
to$(YOUR_MXNET_PACKAGE)
. Thus you can switch among different versions of MXNet quickly. -
Compile dota_kit
sudo apt-get install swig cd ${RoI_ROOT}/dota_kit swig -c++ -python polyiou.i python setup.py build_ext --inplace cd ${RoI_ROOT}/dota_kit/poly_nms_gpu make -j16
-
Prepare script put your original dota data (before split) in path_to_data make sure it looks like
path_to_data/train/images, path_to_data/train/labelTxt, path_to_data/val/images, path_to_data/val/labelTxt, path_to_data/test/images cd ${RoI_ROOT}/prepare_data python prepare_data.py --data_path path_to_data --num_process 32
-
Create soft link
cd ${RoI_ROOT} mkdir -p data cd data ln -s path_to_data dota_1024
We provide trained convnet models.
- To use the demo with our pre-trained RoI Transformer models for DOTA, please download manually from Google Drive, or BaiduYun (Extraction code: fucc)
and put it under the following folder.
Make sure it look like this:
./output/rcnn/DOTA/resnet_v1_101_dota_RoITransformer_trainval_rcnn_end2end/train/rcnn_dota-0040.params ./output/fpn/DOTA/resnet_v1_101_dota_rotbox_light_head_RoITransformer_trainval_fpn_end2end/train/fpn_DOTA_oriented-0008.params
cd ${RoI_ROOT}
-
training Please download ImageNet-pretrained ResNet-v1-101 model manually from OneDrive, or BaiduYun, or Google drive, and put it under folder
./model
. Make sure it look like this:./model/pretrained_model/resnet_v1_101-0000.params
Start training (we use the Light-head R-CNN + RoI Transformer (without FPN) for example, you may choose other models)
sh train_dota_light_RoITransformer.sh
-
testing
Start testing
sh test_dota_light_RoITransformer.sh
© Microsoft, 2017. Licensed under an MIT license.
If you find RoI Transformer and DOTA data useful in your research, please consider citing:
@inproceedings{ding2019learning,
title={Learning RoI Transformer for Oriented Object Detection in Aerial Images},
author={Ding, Jian and Xue, Nan and Long, Yang and Xia, Gui-Song and Lu, Qikai},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={2849--2858},
year={2019}
}
@inproceedings{xia2018dota,
title={DOTA: A large-scale dataset for object detection in aerial images},
author={Xia, Gui-Song and Bai, Xiang and Ding, Jian and Zhu, Zhen and Belongie, Serge and Luo, Jiebo and Datcu, Mihai and Pelillo, Marcello and Zhang, Liangpei},
booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
pages={3974--3983},
year={2018}
}