Skip to content

Official PyTorch Implementation of Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition, ICCV 2021

Notifications You must be signed in to change notification settings

arunos728/SELFY

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition


SELFY_block


This is the official implementation of the paper "Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition" by H.Kwon, M.Kim, S.Kwak, and M.Cho. For more information, checkout the project website and the paper on arXiv.

Environment:

Anaconda environment setting

git clone https://github.com/arunos728/SELFY.git
cd selfy
conda env create -f environment.yml
conda activate selfy

Installing Correlation sampler

cd Pytorch-Correlation-extension
python setup.py install

# check whether SpatialCorrelationSampler is installed correctly.
python check.py forward
python check.py backward
python checkCorrelationSampler.py

Please check this repo for the detailed instructions.

Dataset preparation

Please refer to TSM repo for the detailed data preparation instructions.

File lists (.txt files in ./data) specify configurations of each video clips (path, #frames, class). We upload our Something-Something-V1 & V2 video file lists in ./data. The path of the file lists should be added into the scripts for training (or testing).

Training & Testing

  • For training SELFYNet on Something-Something, use the following command:
    ./scripts/train_SELFY_Something.sh
  • For testing your trained model on Something-Something, use the following command:
    ./scripts/test_SELFY_Something.sh

Citation

If you use this code or ideas from the paper for your research, please cite our paper:

@inproceedings{kwon2021learning,
  title={Learning self-similarity in space and time as generalized motion for video action recognition},
  author={Kwon, Heeseung and Kim, Manjin and Kwak, Suha and Cho, Minsu},
  booktitle={Proceedings of the IEEE/CVF International Conference on Computer Vision},
  pages={13065--13075},
  year={2021}
}

Contact

Heeseung Kwon(aruno@postech.ac.kr), Manjin Kim(mandos@postech.ac.kr)

Questions can also be left as issues in the repository. We will be happy to answer them.

About

Official PyTorch Implementation of Learning Self-Similarity in Space and Time as Generalized Motion for Video Action Recognition, ICCV 2021

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published