Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Incorrect Imagenet dataset #6

Open
prafullasd opened this issue Jun 8, 2022 · 1 comment
Open

Incorrect Imagenet dataset #6

prafullasd opened this issue Jun 8, 2022 · 1 comment
Assignees
Labels
bug Something isn't working

Comments

@prafullasd
Copy link

I downloaded the Imagenet dataset linked in this repo, and I think the dataset for it (50000 test images with labels, box downsampling) doesn't match the official Imagenet 32x32/64x64 versions used for NLL benchmarks (https://github.com/openai/vdvae/blob/main/setup_imagenet.sh, 49999 test images with no labels, can download from https://academictorrents.com/details/96816a530ee002254d29bf7a61c0c158d3dedc3b). Difference in downsampling method used during pre-processing will make the NLL's not comparable.

@Rayhane-mamah Rayhane-mamah self-assigned this Jun 8, 2022
@Rayhane-mamah Rayhane-mamah added the bug Something isn't working label Jun 8, 2022
@Rayhane-mamah
Copy link
Owner

Hello @prafullasd

Thank you for the interest you show in this work and thanks for reaching out about his issue!

You bring up a very good point. We have had our skepticism about the results achieved on Imagenet datasets in our work when the NLL results were very different from the VDVAE baseline. The confusing part about all of this is that the Imagenet version used in NLL benchmarks used to be hosted on this website (as you pointed out by the VDVAE download script), and it seems that since the update of that website, the downsampled imagenet is now downsampled in a different method (this is the part we missed). To add to that confusion even more, some prior work also seems to use the incorrect Imagenet version.

For completeness and future reference:

NLL reported results on Imagenet will probably change when we use these two versions of the dataset. We will update these metrics as we re-do the experiments (will take some time). The general expectation is that our reported NLL should get closer to the NLL reported by VDVAE, which would make sense and match our findings on FFHQ 5-bits.

While noting this mistake is very important and will ensure full research correctness, it doesn't affect the main contribution of the paper much: "Efficient-VDVAE keeps comparable or better NLL performance with less memory and faster training". Nevertheless, being precise in reporting the results is very much desirable.

Thank you very much for pointing this out and helping improve the quality of our work!
I will keep the issue open until we make our updates.
If you find any other problems, please let us know, we appreciate this a lot!

Rayhane.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants