未验证 提交 4c7d5045 编写于 作者: K Kim Yann 提交者: GitHub

rem cncl in ut & build sh (#52811)

* rem cncl in new test

* rem cncl in build sh

* rem cncl in old test
上级 fa8abeec
此差异已折叠。
......@@ -704,7 +704,6 @@ class TestParallelDyGraphRunnerBase:
args.update_method == "nccl2"
or args.update_method == "bkcl"
or args.update_method == "hccl"
or args.update_method == "cncl"
):
strategy = paddle.distributed.parallel.ParallelStrategy()
strategy.nranks = nranks
......@@ -862,7 +861,6 @@ def runtime_main(test_class):
"nccl2_reduce_layer",
"gloo",
"hccl",
"cncl",
],
)
parser.add_argument('--trainer_id', type=int, required=False, default=0)
......@@ -971,7 +969,6 @@ class TestDistBase(unittest.TestCase):
self._bkcl_mode = False
self._gloo_mode = False # now, support gloo backend
self._hccl_mode = False
self._cncl_mode = False
self._pipeline_mode = False
self._mp_mode = False
self._diff_batch = False
......@@ -1778,14 +1775,6 @@ class TestDistBase(unittest.TestCase):
check_error_log=check_error_log,
log_name=log_name,
)
elif self._cncl_mode:
tr0_losses, tr1_losses = self._run_cluster_nccl2(
model_file,
required_envs,
update_method='cncl',
check_error_log=check_error_log,
log_name=log_name,
)
elif self._pipeline_mode:
tr0_losses, tr1_losses = self._run_pipeline(
model_file, required_envs, check_error_log, log_name=log_name
......
......@@ -93,8 +93,7 @@ cc_test(
DEPS tracer layer prepared_operator mul_op)
if(WITH_NCCL
OR WITH_RCCL
OR WITH_XPU_BKCL
OR WITH_CNCL)
OR WITH_XPU_BKCL)
cc_test(
test_group
SRCS test_group.cc
......
......@@ -77,8 +77,7 @@ void GroupConcatSplit(Place place, size_t size) {
}
if (std::is_same<Place, platform::CUDAPlace>::value) {
#if defined(PADDLE_WITH_NCCL) || defined(PADDLE_WITH_RCCL) || \
defined(PADDLE_WITH_CNCL)
#if defined(PADDLE_WITH_NCCL) || defined(PADDLE_WITH_RCCL)
paddle::memory::Copy(
place, data, cpu_place, value.data(), sizeof(T) * value.size(), 0);
#endif
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册