Skip to content
forked from SRA2/SPELL

Learning Long-Term Spatial-Temporal Graphs for Active Speaker Detection (ECCV 2022)

License

Notifications You must be signed in to change notification settings

ErikValleS/SPELL

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

5 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

SPELL

Learning Long-Term Spatial-Temporal Graphs for Active Speaker Detection (ECCV 2022)
paper | poster | presentation (will be updated soon)

Overview

SPELL is a novel spatial-temporal graph learning framework for active speaker detection (ASD). It can model a minute-long temporal contexts without relying on computationally expensive networks. Through extensive experiments on the AVA-ActiveSpeaker dataset, we demonstrate that learning graph-based representations significantly improves the detection performance thanks to its explicit spatial and temporal structure. Specifically, SPELL outperforms all previous state-of-the-art approaches while requiring significantly lower memory and computation resources.

ActivityNet 2022 (AVA-ActiveSpeaker Challenge)

SPELL achieved 2nd place in the AVA-ActiveSpeaker Challenge at ActivityNet 2022. For the challenge, we used a visual input spanning a longer period of time (23 consecutive face-crops instead of 11). We also found that using a larger channel1 can further boost the performance.
tech report | presentation

Dependency

We used python=3.6, pytorch=1.9.1, and torch-geometric=2.0.3 in our experiments.

Code Usage

  1. Download the audio-visual features and the annotation csv files from Google Drive. The directories should look like as follows:
|-- features
    |-- resnet18-tsm-aug
        |-- train_forward
        |-- val_forward
    |-- resnet50-tsm-aug
        |-- train_forward
        |-- val_forward
|-- csv_files
    |-- ava_activespeaker_train.csv
    |-- ava_activespeaker_val.csv
  1. Run generate_graph.py to create the spatial-temporal graphs from the features:
python generate_graph.py --feature resnet18-tsm-aug

Although this script takes some time to finish in its current form, it can be modified to run in parallel and create the graphs for multiple videos at once. For example, you can change the files variable in line 81 of data_loader.py.

  1. Use train_val.py to train and evaluate the model:
python train_val.py --feature resnet18-tsm-aug

You can change the --feature argument to resnet50-tsm-aug for SPELL with ResNet-50-TSM.

Note

  • We used the official code of Active Speakers in Context (ASC) to extract the audio-visual features (Stage-1). Specifically, we used STE_train.py and STE_forward.py of the ASC repository to train our two-stream ResNet-TSM encoders and extract the audio-visual features. We did not use any other components such as the postprocessing module or the context refinement modules. Please refer to models_stage1_tsm.py and the checkpoints from this link to see how we implanted the TSM into the two-stream ResNets.

Citation

ECCV 2022 paper:

will be updated soon

Technical report for AVA-ActiveSpeaker challenge 2022:

@article{minintel,
  title={Intel Labs at ActivityNet Challenge 2022: SPELL for Long-Term Active Speaker Detection},
  author={Min, Kyle and Roy, Sourya and Tripathi, Subarna and Guha, Tanaya and Majumdar, Somdeb},
  year={2022}
}

About

Learning Long-Term Spatial-Temporal Graphs for Active Speaker Detection (ECCV 2022)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%