-
Notifications
You must be signed in to change notification settings - Fork 13
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to reach FID 1.92 on Imagenet 64 #8
Comments
I got similar results although I tried with QKVAttentionLegacy and XformersAttention and suspect that the issues raised here could be hurting results? Which form of attention did you use? |
I also realised that the inference code doesnt seem to use rejection sampling as shown in the paper. This line seems to show that rejection sampling was ran at a ratio of 10,000 except the code referenced doesn't seem to exist in this repo. There is also this file with no documentation, however it seems the most promising. |
I am on a newer version of NVIDIA GPU, i.e. L40S, so no worry on the hardware support of Attention and xformers. Did you manage to merge the rejection sampling into the evaluation pipeline? |
I ran the classifier rejection code but it seemed to produce similar results so I emailed the authors to ask about the difference in performance. |
Sorry I am a newbie here, do you mind pasting your command to run code/classifier_rejection.py here? I literally have no idea how to do that. CTM claimed to use this sampling strategy to achieve a better result, I am confused how to do that. |
I was able to get the published performance by running |
Ah, I see. That makes perfect sense. By the way, do you still need classifier rejection sampling after changing to 50000 samples? |
No, I just used |
I downloaded the pretrained Imagenet 64 checkpoint and use the provided sampling commands (with slight modification to make it run on my machine)
And I tested it according to the provided instrutions.
But the performance reading is significantly worse.
How exactly could I read FID 1.92 ? Even with pretrained model directly from the author?
The text was updated successfully, but these errors were encountered: