diff --git a/python/paddle/v2/framework/tests/test_nccl_allreduce_op.py b/python/paddle/v2/framework/tests/test_nccl_allreduce_op.py index f79dcd664b20325dd9d222ab9d4dcb3132d5d6cd..0a9163dd55c587ad8c56b4f664c17267ba27325b 100644 --- a/python/paddle/v2/framework/tests/test_nccl_allreduce_op.py +++ b/python/paddle/v2/framework/tests/test_nccl_allreduce_op.py @@ -36,9 +36,6 @@ input_data = [ ] output_data = allreduce(input_data, gpus) -# output_vars = [g_scope.var("Out_"+str(i)).get_tensor() -# for i in range(len(gpus))] - def thread_allreduce_op(thread_id, gpu_id): i = gpu_id @@ -53,9 +50,6 @@ def thread_allreduce_op(thread_id, gpu_id): op = create_op(scope, "ncclAllReduce", inputs, outputs, attrs={}) place = core.GPUPlace(gpus[i]) set_input(scope, op, inputs, place) - # # print scope.find_var("Out").get_tensor() - # # print scope.find_var("X").get_tensor() - print scope.find_var("Communicator").get_communicator() ctx = core.DeviceContext.create(place)