Skip to content

Commit

Permalink
update for py3.6 bug (#46325)
Browse files Browse the repository at this point in the history
  • Loading branch information
wuhuachaocoding authored Sep 21, 2022
1 parent b027652 commit 311eb87
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions python/paddle/incubate/optimizer/distributed_fused_lamb.py
Original file line number Diff line number Diff line change
Expand Up @@ -13,13 +13,13 @@
# limitations under the License.

import os
import paddle
from paddle.fluid import framework, core, layers, unique_name
from paddle.fluid.framework import Variable
from paddle.fluid.clip import ClipGradByGlobalNorm
from paddle.fluid.initializer import Constant
from paddle.fluid.layer_helper import LayerHelper
from paddle.fluid.optimizer import Optimizer
import paddle.distributed as dist
from paddle.distributed.collective import new_group
from paddle.fluid.executor import global_scope
from paddle.fluid.framework import name_scope
Expand Down Expand Up @@ -288,8 +288,8 @@ def _apply_gradients_impl(self, params_grads):

step = self._get_or_create_step()

rank = dist.get_rank()
nranks = dist.get_world_size()
rank = paddle.distributed.get_rank()
nranks = paddle.distributed.get_world_size()
if self._nproc_per_node is None:
nproc_per_node = nranks
else:
Expand Down

0 comments on commit 311eb87

Please sign in to comment.