-
Notifications
You must be signed in to change notification settings - Fork 1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Accelerate refuse to work on balanced_low_0
when GPU 0 is not filled.
#2429
Comments
Hi @xkszltl , could you tell me more about your specific use case and the issue that you are facing ? I don't understand the part about the device map since we do compute it before and we do |
We are loading LLAMA2 7B and Mistral for finetuning in a single node 8 GPU setup. |
The part I don’t understand is what make GPU 0 so special and worth asserting here? |
Hi @xkszltl, this check might be outdated since we added the possibility to fine-tune BNB model with naive PP . I'll let @younesbelkada confirms this point ! |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Not stale |
@SunMarc @younesbelkada |
Hi @xkszltl, sorry for the delay. Would you like to submit a PR to fix this and check that the tests are passing on transformers and accelerate ? Thanks ! |
Seems `balanced_low_0` can leave GPU 0 empty and breaks this check. According to the discussion this check may be outdated. Resolve huggingface#2429
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Not stale |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Closing this since this issue should be solved by this PR |
System Info
accelerate/src/accelerate/accelerator.py
Lines 1324 to 1333 in 6f05bbd
This part would throw if model is loaded by
transformers.AutoModelForCausalLM.from_pretrained(..., device_map="balanced_low_0", ...)
, because GPU 0 may be left completely unused.This doesn't seem to be a good behavior as there's no way to tell what's the "first device" without computing the device map first.
The text was updated successfully, but these errors were encountered: