-
Notifications
You must be signed in to change notification settings - Fork 30
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Support for multiple GPUs #19
Comments
Hello, We removed it from the public repo to make the code a bit easier to read and support. You can likely do something similar to the nvdiffrec mGPU setup. |
Hello again, I have a working first implementation but I am encountering some problems as far as GPU memory is concerned. For example, I still get CUDA out of memory errors when I increase the batch size by one despite effectively having x4 memory. This leads me to suspect that some processes are shared between the GPUs. Maybe you have some insights to help me here? I have also noticed that |
Hey were you able to get the multi-gpu setup working? |
Yeah partially, I had to do a few hacky things in the data processing to get it to work. But at the end I could process a batch across multiple gpus. |
I tried with distributed data parallel and somewhat changing to pytorvch lightning but getting seg fault on pretty low res images and batch sizes. How did you resolve it? It would be nice to discuss further |
I never got a segfault error. How are you trying to implement the parallelism? I think you can only expect to divide the batches across the gpus. Here is briefly what I have done:
|
@Selozhd thanks! Can you share the code for reference~~ |
I am planning to run the model on multiple GPUs. However, Looking at the way
optimize_mesh()
is written it is not immediately clear how to implement it. In nvdiffrec, there used to be multiple gpu support implemented through aTrainer
class.Is any particular reason why you removed it?
The text was updated successfully, but these errors were encountered: