提交 026c61c0 编写于 作者: D Dong Zhihong

"fix allreduce python test"

上级 1e8474b9
...@@ -36,9 +36,6 @@ input_data = [ ...@@ -36,9 +36,6 @@ input_data = [
] ]
output_data = allreduce(input_data, gpus) output_data = allreduce(input_data, gpus)
# output_vars = [g_scope.var("Out_"+str(i)).get_tensor()
# for i in range(len(gpus))]
def thread_allreduce_op(thread_id, gpu_id): def thread_allreduce_op(thread_id, gpu_id):
i = gpu_id i = gpu_id
...@@ -53,9 +50,6 @@ def thread_allreduce_op(thread_id, gpu_id): ...@@ -53,9 +50,6 @@ def thread_allreduce_op(thread_id, gpu_id):
op = create_op(scope, "ncclAllReduce", inputs, outputs, attrs={}) op = create_op(scope, "ncclAllReduce", inputs, outputs, attrs={})
place = core.GPUPlace(gpus[i]) place = core.GPUPlace(gpus[i])
set_input(scope, op, inputs, place) set_input(scope, op, inputs, place)
# # print scope.find_var("Out").get_tensor()
# # print scope.find_var("X").get_tensor()
print scope.find_var("Communicator").get_communicator()
ctx = core.DeviceContext.create(place) ctx = core.DeviceContext.create(place)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册