Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CUDA out of memory #207

Open
DawnJYe opened this issue Jan 4, 2025 · 2 comments
Open

CUDA out of memory #207

DawnJYe opened this issue Jan 4, 2025 · 2 comments

Comments

@DawnJYe
Copy link

DawnJYe commented Jan 4, 2025

Does running this model require a lot of GPU memory? Will using an A100 GPU still result in an out-of-memory error?

@DawnJYe
Copy link
Author

DawnJYe commented Jan 6, 2025

Hello, we tried to solve the issue.

This is what we did:

Modify the Graphormer3D model to use gradient checkpointing and reduce batch size to decrease GPU memory usage.

You can review changes in this commit: trinks-slam818@15984ac.

Caution

Disclaimer: The concept of solution was created by AI and you should never copy paste this code before you check the correctness of generated code. Solution might not be complete, you should use this code as an inspiration only.

Latta AI seeks to solve problems in open source projects as part of its mission to support developers around the world. Learn more about our mission at https://latta.ai/ourmission . If you no longer want Latta AI to attempt solving issues on your repository, you can block this account.

I don't use graphormer3D and I just use base graphormer.It seems that the batchsize needs to be reduced to very small for normal training.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants
@DawnJYe and others