-
Notifications
You must be signed in to change notification settings - Fork 179
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model will be loaded on different devices when using multiple gpus. #67
Comments
May I ask in which line of inference did this error occur? |
Sorry for the delay. Here is one error message:
|
You might also want to try setting device_map=auto in your model_args when you do multi-processing
|
Setting device_map to auto didn't do the trick. Here's my command:
I noticed one difference between evaluation using Line 104 appears to be here. |
Sorry, my bad Should set |
Thanks. Now it works! |
It appears that models will be loaded on different gpus when setting
num_processes
to more than one, which will cause error:RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!
Here's my command to launch:
I found a temporary fix by installing previous version:
pip install git+https://github.com/EvolvingLMMs-Lab/lmms-eval.git@bf4c78b7e405e2ca29bf76f579371382fec3dd02
and in this version multi-gpu inference works fine.
The text was updated successfully, but these errors were encountered: