Official PyTorch code for CVPR'22 paper "Remember Intentions: Retrospective-Memory-based Trajectory Prediction".
Abstract: To realize trajectory prediction, most previous methods adopt the parameter-based approach, which encodes all the seen past-future instance pairs into model parameters. However, in this way, the model parameters come from all seen instances, which means a huge amount of irrelevant seen instances might also involve in predicting the current situation, disturbing the performance. To provide a more explicit link between the current situation and the seen instances, we imitate the mechanism of retrospective memory in neuropsychology and propose MemoNet, an instance-based approach that predicts the movement intentions of agents by looking for similar scenarios in the training data. In MemoNet, we design a pair of memory banks to explicitly store representative instances in the training set, acting as prefrontal cortex in the neural system, and a trainable memory addresser to adaptively search a current situation with similar instances in the memory bank, acting like basal ganglia. During prediction, MemoNet recalls previous memory by using the memory addresser to index related instances in the memory bank. We further propose a two-step trajectory prediction system, where the first step is to leverage MemoNet to predict the destination and the second step is to fulfill the whole trajectory according to the predicted destinations. Experiments show that the proposed MemoNet improves the FDE by 20.3%/10.2%/28.3% from the previous best method on SDD/ETH-UCY/NBA datasets. Experiments also show that our MemoNet has the ability to trace back to specific instances during prediction, promoting more interpretability.
We give an example of trajectories predicted by our model and the corresponding ground truth as following:
Below is an example of prediction interpretability where the first column stands for the current agent. The last three columns stand for the memory instances found by the current agent.
You can find the code and the instructions in the ETH folder.
- Tested OS: Linux / RTX 3090
- Python == 3.7.9
- PyTorch == 1.7.1+cu110
Install the dependencies from the requirements.txt
:
pip install -r requirements.txt
We provide a complete set of pre-trained models including:
- intention encoder-decoder:
- learnable addresser:
- generated memory bank:
- fulfillment encoder-decoder:
You can download the pretrained models/data from here.
After the prepartion work, the whole project should has the following structure:
./MemoNet
├── ReadMe.md
├── data # datasets
│ ├── test_all_4096_0_100.pickle
│ └── train_all_512_0_100.pickle
├── models # core models
│ ├── layer_utils.py
│ ├── model_AIO.py
│ └── ...
├── requirements.txt
├── run.sh
├── sddloader.py # sdd dataloader
├── test_MemoNet.py # testing code
├── train_MemoNet.py # training code
├── trainer # core operations to train the model
│ ├── evaluations.py
│ ├── test_final_trajectory.py
│ └── trainer_AIO.py
└── training # saved models/memory banks
├── saved_memory
│ ├── sdd_social_filter_fut.pt
│ ├── sdd_social_filter_past.pt
│ └── sdd_social_part_traj.pt
├── training_ae
│ └── model_encdec
├── training_selector
│ ├── model_selector
│ └── model_selector_warm_up
└── training_trajectory
└── model_encdec_trajectory
Important configurations.
--mode
: verify the current training mode,--model_ae
: pretrained model path,--info
: path name to store the models,--gpu
: number of devices to run the codes,
Training commands.
bash run.sh
To get the reported results, following
python test_MemoNet.py --reproduce True --info reproduce --gpu 0
And the code will output:
./training/training_trajectory/model_encdec_trajectory
Test FDE_48s: 12.659514427185059 ------ Test ADE: 8.563031196594238
----------------------------------------------------------------------------------------------------
Thanks for the framework provided by Marchetz/MANTRA-CVPR20
, which is source code of the published work MANTRA in CVPR-2020. The github repo is MANTRA code. We borrow the framework and interface from the code.
We also thank for the pre-processed data provided by the works of PECNet (paper,code).
If you use this code, please cite our paper:
@InProceedings{MemoNet_2022_CVPR,
author = {Xu, Chenxin and Mao, Weibo and Zhang, Wenjun and Chen, Siheng},
title = {Remember Intentions: Retrospective-Memory-based Trajectory Prediction},
booktitle = {The IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
year = {2022}
}