Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about loss selection #6

Open
Yu-Fangxu opened this issue Apr 16, 2024 · 1 comment
Open

Question about loss selection #6

Yu-Fangxu opened this issue Apr 16, 2024 · 1 comment

Comments

@Yu-Fangxu
Copy link

Hi the authors,
I learned a lot from your wonderful work. I want to ask if you have ever tried Trajectory Balance(TB) as the learning objective. I used to run your code and found that the loss starts from 10^3 to 10^4, which is large. When considering TB, the loss should not be that large, because XYZ should be consistent. thus leading to a large P(XZY).
So did you try TB at first? If there is any default when training with TB?

Thanks!

@MJ10
Copy link
Contributor

MJ10 commented Apr 18, 2024

Hi @Yu-Fangxu, we did try trajectory balance at some point early on in the project but haven't tried it since. All the experiments were run with the modified SubTB loss, so unfortunately I can't help with good defaults. Regarding the loss - if you are initializing new LoRA weights, the initial loss can be quite high - in some other projects we have observed that for short sequences TB does work well. Hope this helps!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants