Official implementation of: "Video Prediction at Multiple Spatio-Temporal Scales with Hierarchical Recurrent Networks" by Villar-Corrales et al. [Paper] [Project]
You can easily install all required packages and avoiding dependency issues by installing the conda
environment file included in the repository. To do so, run the following command from the terminal:
$ conda env create -f environment.yml
$ conda activate MSPred
To obtain the datasets, follow the steps:
-
MovingMNIST: This dataset is generated on the fly using images from the popular MNIST dataset. The original MNIST is automatically downloaded using
torchvision
. -
KTH-Actions: You can download and preprocess this dataset using the bash scripts in the
resources
directory:
cd resources
./download_kth.sh $DATA_PATH
./convert_kth.sh $DATA_PATH
In case the download scripts are not working, please download the KTH-Actions dataset from this shared directory
- SynpickVP: You can download this dataset from this shared directory
The following tree diagram displays the detailed directory structure of the project. Directory names and paths can be modified in the CONFIG File (/src/CONFIG.py
).
MSPred
├── datasets/
| ├── KTH-64/
| ├── SYNPICK/
| └── ...
├── experiments/
| ├── exp_mmnist/
| └── exp_kth/
├── resources/
| └── ...
├── src/
| └── ...
├── environment.yml
└── README.md
The following table contains links to pretrained MSPred models on the MovingMNIST, KTH-Actions and SynpickVP datasets. These models can be used to reproduce the results from our paper.
Additionally, we include links to training logs and Tensorboards, which show the training and validation progress of our model.
Dataset | Model & Config | Logs |
---|---|---|
Moving MNIST | Experiment | tboard |
KTH-Actions | Experiment | tboard |
SynpickVP | Experiment | tboard |
Follow this section for a quick guide on how to get started with this repository.
Our codebase is based is structured in an experiment-oriented manner. Experiments contains an experiment_params.json with all the hyper-parameters and specifications needed to train and evaluate a model.
Creating an experiment automatically generates a directory in the specified EXP_DIRECTORY, containing a JSON file with the experiment parameters and sub-directories for the models, plot, and Tensorboard logs.
$ python src/01_create_experiment.py [-h] -d EXP_DIRECTORY [--name NAME] [--config CONFIG]
optional arguments:
-h, --help show this help message and exit
-d EXP_DIRECTORY, --exp_directory EXP_DIRECTORY Directory where the experiment folder will be created
--name NAME Name to give to the experiment
--config CONFIG Name of the predetermined 'config' to use
Once the experiment is initialized and the experiment parameters are set to the desired values, a model can be trained following command:
$ CUDA_VISIBLE_DEVICES=0 python src/02_train.py [-h] -d EXP_DIRECTORY [--checkpoint CHECKPOINT] [--resume_training]
optional arguments:
-h, --help show this help message and exit
-d EXP_DIRECTORY, --exp_directory EXP_DIRECTORY Path to the experiment directory
--checkpoint CHECKPOINT Checkpoint with pretrained parameters to load
--resume_training For resuming training
Model checkpoints, which are saved regularly during training, can be evaluated using the following command:
$ CUDA_VISIBLE_DEVICES=0 python src/3_evaluate.py [-h] -d EXP_DIRECTORY [--checkpoint CHECKPOINT]
optional arguments:
-h, --help show this help message and exit
-d EXP_DIRECTORY, --exp_directory EXP_DIRECTORY Path to the experiment directory
--checkpoint CHECKPOINT Checkpoint with pretrained parameters to load
$ python src/01_create_experiment.py -d TestExp --name exp_mmnist --config mmnist
$ CUDA_VISIBLE_DEVICES=0 python src/02_train.py -d experiments/TestExp/exp_mmnist/
$ CUDA_VISIBLE_DEVICES=0 python src/03_evaluate.py -d experiments/TestExp/exp_mmnist/ --checkpoint checkpoint_epoch_final.pth
This repository is maintained by Angel Villar-Corrales,
Please consider citing our paper if you find our work or our repository helpful.
@inproceedings{villar2022MSPred,
title={MSPred: Video Prediction at Multiple Spatio-Temporal Scales with Hierarchical Recurrent Networks},
author={Villar-Corrales, Angel and Karapetyan, Ani and Boltres, Andreas and Behnke, Sven},
booktitle={British Machine Vision Conference (BMVC)},
year={2022}
}
In case of any questions or problems regarding the project or repository, do not hesitate to contact the authors at villar@ais.uni-bonn.de.