[Paper] (https://arxiv.org/abs/2307.01985)
This paper has been accepted by Neurocomputing.(https://doi.org/10.1016/j.neucom.2024.128044)
This repo contains code for the method introduced in the paper: (https://github.com/cofly2014/TSA-MLT)
[Task-Specific Alignment and Multiple-level Transformer for Few-Shot Action Recognition]
##TODO
##TODO
We used https://github.com/ffmpbgrnn/CMN for Kinetics and SSv2, which are provided by the authors of the authors of CMN (Zhu and Yang, ECCV 2018). We also used the split from OTAM (Cao et al. CVPR 2020) for SSv2, and splits from ARN (Zhang et al. ECCV 2020) for HMDB and UCF. These are all the in the splits folder.
If you use this code/method or find it helpful, please cite:
We based our code on TRX Temporal-Relational CrossTransformers for Few-Shot Action Recognition CVPR2021 (logging, training, evaluation etc.). and the TA2N: Two-Stage Action Alignment Network for Few-Shot Action Recognition AAAI2022.(modify and improve the tmm) and the We use torch_videovision for video transforms. We took inspiration from the image-based CrossTransformer