-
Notifications
You must be signed in to change notification settings - Fork 91
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feat] Implementing GLOP #253
Conversation
Co-authored-by: Chuanbo Hua <cbhua@kaist.ac.kr>
Co-authored-by: Chuanbo Hua <cbhua@kaist.ac.kr>
Despite all these efforts, the current version of the GLOP model still refuses to learn 😞. However, I've obtained a working checkpoint in Oct. 2024. Since then, I’ve tried numerous times to replicate the success, but no luck so far. It seems like getting it to learn requires a bit of magic 🤔. At least this proves that the basic code logic is working as expected. In theory, this checkpoint should work with this version of code. I’ll upload the checkpoint soon after some tests! |
Great job @Furffico ! I think it may also be a problem of the environment in reproducing the main paper results, as seen here. Perhaps downgrading to e.g. PyTorch 2.2 / changing to FP32 could do the trick? This said, given the code is indeed correct per se if it's indeed a matter of environment as it seems to be I would be for merging! |
Description
Following #182, this PR includes the implementation of Global and Local Optimization Policies (GLOP) (Ye et al., 2023), along with these new features for reproducing the results:
Types of changes
Checklist
CC: @henry-yeh @fedebotu