-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
mutil gpus #507
Comments
Did you find a solution for this? |
I have found a solution to this. I leverage Pytorch distributed package with Nvidia Apex to train on 4 rtx 2080ti with a batch size of 80. |
How to do that for your saying 'leverage Pytorch distributed package with Nvidia Apex to train on 4 rtx 2080ti with a batch size of 80' Is that easy or not, Thank you! |
@jxhno1 it is definitely not very hard to do. It is just time-consuming for first-timer. You will have to dig deep into PyTorch Cuda API functionalities. But once you are able to deploy one model, you can do for all of them. It is something everyone should pick up since it can be applied generally. |
can you solute multi gpu for training?Thanks! |
Duplicate of #520 |
Has anyone solved the mutil-gpus traing probem?
The text was updated successfully, but these errors were encountered: