未验证 提交 4465ba27 编写于 作者: H Haohongxiang 提交者: GitHub

rename fw_bw func name of interleave pp (#47571) (#47862)

上级 3a6cc57c
...@@ -537,7 +537,7 @@ class PipelineParallelWithInterleave(PipelineParallel): ...@@ -537,7 +537,7 @@ class PipelineParallelWithInterleave(PipelineParallel):
return input_tensor_grad return input_tensor_grad
def interleave_pipeline( def forward_backward_pipeline(
self, data, scaler, forward_only=False, compute_loss=True self, data, scaler, forward_only=False, compute_loss=True
): ):
# use interleave scheduling strategy. # use interleave scheduling strategy.
...@@ -766,7 +766,7 @@ class PipelineParallelWithInterleave(PipelineParallel): ...@@ -766,7 +766,7 @@ class PipelineParallelWithInterleave(PipelineParallel):
def train_batch(self, data, optimizer, lr_scheduler=None, scaler=None): def train_batch(self, data, optimizer, lr_scheduler=None, scaler=None):
data = self._prepare_training(data, optimizer, lr_scheduler) data = self._prepare_training(data, optimizer, lr_scheduler)
# interleave scheduler for pipeline parallel # interleave scheduler for pipeline parallel
train_loss = self.interleave_pipeline(data, scaler) train_loss = self.forward_backward_pipeline(data, scaler)
# optimizer # optimizer
with paddle.amp.auto_cast(enable=False): with paddle.amp.auto_cast(enable=False):
...@@ -781,4 +781,4 @@ class PipelineParallelWithInterleave(PipelineParallel): ...@@ -781,4 +781,4 @@ class PipelineParallelWithInterleave(PipelineParallel):
self._layers.eval() self._layers.eval()
self._compute_loss = compute_loss self._compute_loss = compute_loss
return self.interleave_pipeline(data, None, forward_only=True) return self.forward_backward_pipeline(data, None, forward_only=True)
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册