-
Notifications
You must be signed in to change notification settings - Fork 27.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Beam search fails when using model parallelism #9200
Comments
As the trace suggests, the error seem to come from the I just tried to modify line 229 in
which seems to work. |
FWIW, this fix doesn't currently work for T5, as the fix to |
@OyvindTafjord - would you mind opening a new PR for it? :-) |
Environment info
transformers
version: 4.1.1model.parallelize()
Who can help
@LysandreJik
@alexorona
Information
Model I am using (Bert, XLNet ...): GPT2
The problem arises when using:
The tasks I am working on is:
To reproduce
The recent (and awesome!) model
parallelize()
doesn't seem to work with beam search decoding at the moment. The behavior can be reproduced on the officialhuggingface/transformers-pytorch-gpu:4.1.1
docker image by running the following (on a machine with multiple GPUs):This raises the following error:
Expected behavior
The expected behavior is to not raise an error, but instead correctly return the beam search decoding.
The text was updated successfully, but these errors were encountered: