diff --git a/python/paddle/distributed/fleet/meta_optimizers/tensor_parallel_optimizer.py b/python/paddle/distributed/fleet/meta_optimizers/tensor_parallel_optimizer.py index 5fbec7da0b5edfbe9b7b84bd8e41f1eb1f1ae098..9d099a2af24facc25d36ecbfcb287b422858293e 100644 --- a/python/paddle/distributed/fleet/meta_optimizers/tensor_parallel_optimizer.py +++ b/python/paddle/distributed/fleet/meta_optimizers/tensor_parallel_optimizer.py @@ -124,7 +124,7 @@ class TensorParallelOptimizer(MetaOptimizerBase): collective_helper._init_communicator( self.startup_program, self.current_endpoint, self.mp_endpoints, self.mp_rank, self.mp_ring_id, True, self.global_ring_id, True) - #self._broadcast_params(self.mp_ring_id, mp_mode=True) + self._broadcast_params(self.mp_ring_id, mp_mode=True) # Create dp rings if self.nranks > self.mp_degree: