-
Notifications
You must be signed in to change notification settings - Fork 18
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
A confusing question in transformer training #15
Comments
Hi @forever-rz , Thanks for your interests. |
@liuqk3 Thanks for your help! Can you tell me how should I use just one codebook ? |
My configuration is shown below:
model:
|
|
Hi, @forever-rz . Sorry for the delayed reply.
|
Hi, @liuqk3 .I'm sorry that I temporarily put aside the experiment because I couldn't figure out the cause of this problem. Today, I carefully checked the previous experiment and found the relevant data of the three questions you raised as follows. PLACES2 3 The loss of training transformer based on my P-VQVAE is as follows (ImageNet is too big, so it is given up) |
It seems to me that the reconstruction results are not bad, so I don't understand why transformer's training results are so wrong. Although I made some changes to P-VQVAE, transformer has not changed at all. |
@forever-rz , |
Hi, do you have any updates on this issue? I have also encountered same problem with a custom dataset, reconstruction results are much better than completed results. |
I also encountered the problem of good reconstruction effect but poor generation effect in a similar task. I saw that our loss curves are basically the same. Did you solve this problem in the end? |
Thanks for your contribution,but there is a problem when i train it on FFHQ.Once the ratio of mask is larger, it seems that only part of the completed result is repaired, the part that is not repaired stays black and no new content seems to be generated. Is this normal?



eg1: First on the left, second on the left remain some strange black regions .(epoch:12)
mask_input
completed
reconstruction
The text was updated successfully, but these errors were encountered: