-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
You can't train a model that has been loaded in 8-bit precision on multiple devices #414
Comments
I got the same error, my model is initialized as follows: model = AutoModelForSeq2SeqLM.from_pretrained(
args.model_name_or_path,
config=config,
load_in_8bit=True,
device_map="auto"
) It seems that |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. |
AFAIK one cannot backpropagate gradients through an 8bit model. The idea is to load the base model in 8bit, and the additional LoRA parameters in higher precision. During finetuning, only the higher-precision LoRA parameters are updated. To achieve this, we use However, with that approach, I also get
😐 |
This should have been fixed in huggingface/accelerate#1523 pip install --upgrade accelerate Feel free to re-open the issue if you don't think so |
@younesbelkada Great, thanks! I will try with the latest version. |
Hi,
I was trying to tune Whisper using the script shared in the examples (this one here)
while using multiple GPUs (4), however that throws an error saying:
You can't train a model that has been loaded in 8-bit precision on multiple devices
Is it supported to train a 8bit whisper model on multiple GPUs?
I note that the using 1GPU works fine using the same script. Am I missing something?
The full error message:
The text was updated successfully, but these errors were encountered: