-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
LoRA Fine Tuning Crash at FlashAttention Issue #1828
Comments
cc @BabyChouSr |
me too |
flash attention might not work for your gpu. Head dim > 64 backward requires A100 or H100. You can consider commenting out the flash attention part |
I tried using an A100 as well and I got the same error. |
Same error observed. I found it could train by removing |
may try using xformer. |
Try #2126 |
I tried running this command:
On runpod.io, but I keep getting:
My setup instructions were as follows:
huggyllama/llama-7b
... is this the right HF model to use?Are there any plans to fix this?
The text was updated successfully, but these errors were encountered: