This is the official implementation of the paper "Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition" by H.Kwon, M.Kim, S.Kwak, and M.Cho. For more information, checkout the project website and the paper on arXiv.
- Cuda: 9.0
- gcc: 7.3.0
- Python 3.6.8
- PyTorch 1.0.1
- TorchVison: 0.2.2
- Spatial Correlation Sampler
- Others: environment.yml
git clone https://github.com/arunos728/SELFY.git
cd selfy
conda env create -f environment.yml
conda activate selfy
cd Pytorch-Correlation-extension
python setup.py install
# check whether SpatialCorrelationSampler is installed correctly.
python check.py forward
python check.py backward
python checkCorrelationSampler.py
Please check this repo for the detailed instructions.
Please refer to TSM repo for the detailed data preparation instructions.
File lists (.txt files in ./data) specify configurations of each video clips (path, #frames, class). We upload our Something-Something-V1 & V2 video file lists in ./data. The path of the file lists should be added into the scripts for training (or testing).
- For training SELFYNet on Something-Something, use the following command:
./scripts/train_SELFY_Something.sh
- For testing your trained model on Something-Something, use the following command:
./scripts/test_SELFY_Something.sh
If you use this code or ideas from the paper for your research, please cite our paper:
@inproceedings{kwon2021learning,
title={Learning self-similarity in space and time as generalized motion for video action recognition},
author={Kwon, Heeseung and Kim, Manjin and Kwak, Suha and Cho, Minsu},
booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
pages={13065--13075},
year={2021}
}
Heeseung Kwon(aruno@postech.ac.kr), Manjin Kim(mandos@postech.ac.kr)
Questions can also be left as issues in the repository. We will be happy to answer them.