We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent d45bfa1 commit b725c26Copy full SHA for b725c26
vllm/distributed/parallel_state.py
@@ -1013,6 +1013,7 @@ def initialize_model_parallel(
1013
parallelism.
1014
pipeline_model_parallel_size: number of GPUs used for pipeline model
1015
1016
+ backend: name of torch distributed communication backend.
1017
1018
Let's say we have a total of 8 GPUs denoted by g0 ... g7 and we
1019
use 2 GPUs to parallelize the model tensor, and 4 GPUs to parallelize
0 commit comments