How to use pytorch-lightning distributed training without SLURM? #1334
-
❓How to use pytorch-lightning distributed training without SLURM?Couldn't find anywhere a single note or tutorial on this. For example I have just 2 node with 4 GPUs on each. Is this possible to train pytorch-lightning script in this setup and if so how? |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments
-
you can configure your own environment variables and do your own setup. Just override (corrected) |
Beta Was this translation helpful? Give feedback.
-
you also need to override |
Beta Was this translation helpful? Give feedback.
you can configure your own environment variables and do your own setup.
Just override
LightningModule.init_ddp_connection
https://pytorch-lightning.readthedocs.io/en/latest/lightning-module.html#lightningmodule-class
(corrected)