We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 920a9c2 commit 61b7785Copy full SHA for 61b7785
torchtune/training/_distributed.py
@@ -674,7 +674,7 @@ def shard_model(
674
675
if num_layers_sharded == 0:
676
raise ValueError(
677
- "No layer modules were sharded. Please check if shard conditions is working as expected."
+ "No layer modules were sharded. Please check if shard conditions are working as expected."
678
)
679
680
# Finally shard the entire model to account for any stragglers
0 commit comments