-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
VAE Config Files #27
Comments
The VAE used in our paper is a VQ-f4 VAE (VQGAN). The original LDM repo only supports KL VAE training, so we used the codebase of taming-transformers. We fine-tuned the VAE for 50k iterations on 3 GPUs, with bs=12 per GPU. Diffusers library now supports training a VQ-VAE: https://github.com/huggingface/diffusers/tree/main/examples/vqgan You might want to track image quality metrics like PSNR / SSIM during training. |
What about the discriminator hyper-parameters and the codebook weight for the LPIPS loss? |
On second thought, would it be possible to share the config file you used for this phase? I appreciate all the help! |
I'm interested in fine-tuning a VAE and was wondering the config files you used for the final vq-f4 model in your paper?
I see the existing config files under the autoencoder directory, but they are missing some hyper-parameters like number of training epochs.
Please advise.
The text was updated successfully, but these errors were encountered: