Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Can you please share the steps for reproducing video summarization results on TVSum and SumMe datasets? #2

Open
surajkothawade opened this issue Apr 2, 2019 · 1 comment

Comments

@surajkothawade
Copy link

surajkothawade commented Apr 2, 2019

It seems that the given code is specifically for the mpii dataset.
It would be great if you could share some insights on replicating numbers for other tasks mentioned in the paper.

@GilLevi
Copy link
Collaborator

GilLevi commented Apr 13, 2019

Sure. In order to reproduce our results, you will need to extract every 10th frame from the videos, from each such frame get a vgg19 representation (using the penultimate layer) and then calculate rnn-fv for every sequence of consecutive N frames (where N is a hyperparameter, we checked 4,8 and 16). Once you have those representations, you can split to train/validation and test following each benchmark protocol and use the code in this repository.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants