Skip to content

Commit

Permalink
[pre-commit.ci] auto fixes from pre-commit.com hooks
Browse files Browse the repository at this point in the history
for more information, see https://pre-commit.ci
  • Loading branch information
pre-commit-ci[bot] committed Apr 4, 2023
1 parent 895a2de commit 988532b
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 4 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -335,11 +335,10 @@ def training_step(self, dataloader_iter, batch_idx):
if self.initialize_ub:
input_shape = [
self.cfg.get('encoder_seq_length') * self.cfg.get('micro_batch_size'),
self.cfg.get('hidden_size')
self.cfg.get('hidden_size'),
]
te_module.pre_init_ub(
shape=input_shape,
dtype=torch.uint8 if self.cfg.get('fp8') else torch.bfloat16,
shape=input_shape, dtype=torch.uint8 if self.cfg.get('fp8') else torch.bfloat16,
)
self.initialize_ub = False

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -779,7 +779,7 @@ def __init__(
layer_type: str = "encoder",
drop_path_rate: float = 0,
use_emha: bool = False,
ub_tp_comm_overlap: bool =False,
ub_tp_comm_overlap: bool = False,
autocast_dtype: Any = 16,
zero_centered_gamma: bool = False,
) -> None:
Expand Down

0 comments on commit 988532b

Please sign in to comment.