You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Running with 4xA30 GPUs with 24G each one, while performing the inference I reached out of memory during decoding latents.
Multi-gpu does not include tiled decoding but I think it should so that less VRAM is needed.
I suppose what we should really do is calculate the total amount of VRAM, and then use that to figure out whether or not we should enable tiled decoding.
Same here, running on 2xA100 with 40GB vram each.
Now I got another issue:
torch.distributed.DistBackendError: NCCL error in: ../torch/csrc/distributed/c10d/NCCLUtils.hpp:317, unhandled cuda error (run with NCCL_DEBUG=INFO for details), NCCL version 2.21.5
ncclUnhandledCudaError: Call to CUDA functionfailed.
Last error:
Cuda failure 2 'out of memory'
Running with 4xA30 GPUs with 24G each one, while performing the inference I reached out of memory during decoding latents.
Multi-gpu does not include tiled decoding but I think it should so that less VRAM is needed.
I have hardcoded the following:
mochi/src/genmo/mochi_preview/pipelines.py
Line 677 in f64ffe6
to:
And doing it I've been able to generate the video.
The text was updated successfully, but these errors were encountered: