Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
5200c657
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
5200c657
编写于
10月 25, 2017
作者:
D
Dong Zhihong
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
"move Tensor to LoDTensor"
上级
63fb41b3
变更
4
显示空白变更内容
内联
并排
Showing
4 changed file
with
186 addition
and
105 deletion
+186
-105
paddle/operators/nccl_op.cc
paddle/operators/nccl_op.cc
+7
-0
paddle/operators/nccl_op.cu
paddle/operators/nccl_op.cu
+17
-3
paddle/operators/nccl_op.h
paddle/operators/nccl_op.h
+0
-50
paddle/operators/nccl_op_test.cu
paddle/operators/nccl_op_test.cu
+162
-52
未找到文件。
paddle/operators/nccl_op.cc
浏览文件 @
5200c657
...
@@ -74,8 +74,15 @@ class NCCLAllReduceOp : public framework::OperatorWithKernel {
...
@@ -74,8 +74,15 @@ class NCCLAllReduceOp : public framework::OperatorWithKernel {
// reduction == "ncclMin" || reduction == "ncclMax"),
// reduction == "ncclMin" || reduction == "ncclMax"),
// "invalid reduction.");
// "invalid reduction.");
// auto in_dim = x_dims[0];
ctx
->
SetOutputsDim
(
"Out"
,
x_dims
);
ctx
->
SetOutputsDim
(
"Out"
,
x_dims
);
ctx
->
ShareLoD
(
"X"
,
/*->*/
"Out"
);
ctx
->
ShareLoD
(
"X"
,
/*->*/
"Out"
);
size_t
N
=
x_dims
.
size
();
auto
out_dims
=
ctx
->
GetOutputsDim
(
"Out"
);
for
(
size_t
i
=
0
;
i
<
N
;
++
i
)
{
VLOG
(
1
)
<<
" inference (X) "
<<
framework
::
product
(
x_dims
[
i
])
<<
" (Out)"
<<
framework
::
product
(
out_dims
[
i
]);
}
}
}
};
};
...
...
paddle/operators/nccl_op.cu
浏览文件 @
5200c657
...
@@ -12,6 +12,7 @@ limitations under the License. */
...
@@ -12,6 +12,7 @@ limitations under the License. */
#define EIGEN_USE_GPU
#define EIGEN_USE_GPU
#include <functional>
#include <functional>
#include "paddle/framework/lod_tensor.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/op_registry.h"
#include "paddle/operators/nccl/nccl_gpu_common.h"
#include "paddle/operators/nccl/nccl_gpu_common.h"
...
@@ -20,6 +21,7 @@ namespace operators {
...
@@ -20,6 +21,7 @@ namespace operators {
using
framework
::
Tensor
;
using
framework
::
Tensor
;
using
platform
::
Communicator
;
using
platform
::
Communicator
;
using
framework
::
LoDTensor
;
template
<
typename
Type
>
template
<
typename
Type
>
class
NCCLTypeWrapper
;
class
NCCLTypeWrapper
;
...
@@ -43,8 +45,8 @@ class NCCLAllReduceKernel : public framework::OpKernel<T> {
...
@@ -43,8 +45,8 @@ class NCCLAllReduceKernel : public framework::OpKernel<T> {
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
"This kernel only runs on GPU device."
);
"This kernel only runs on GPU device."
);
auto
ins
=
ctx
.
MultiInput
<
Tensor
>
(
"X"
);
auto
ins
=
ctx
.
MultiInput
<
LoD
Tensor
>
(
"X"
);
auto
outs
=
ctx
.
MultiOutput
<
Tensor
>
(
"Out"
);
auto
outs
=
ctx
.
MultiOutput
<
LoD
Tensor
>
(
"Out"
);
auto
*
comm
=
ctx
.
Input
<
Communicator
>
(
"Communicator"
);
auto
*
comm
=
ctx
.
Input
<
Communicator
>
(
"Communicator"
);
...
@@ -56,12 +58,24 @@ class NCCLAllReduceKernel : public framework::OpKernel<T> {
...
@@ -56,12 +58,24 @@ class NCCLAllReduceKernel : public framework::OpKernel<T> {
boost
::
get
<
platform
::
GPUPlace
>
(
ctx
.
GetPlace
()).
GetDeviceId
();
boost
::
get
<
platform
::
GPUPlace
>
(
ctx
.
GetPlace
()).
GetDeviceId
();
int
idx
=
comm
->
GetCommId
(
device_id
);
int
idx
=
comm
->
GetCommId
(
device_id
);
size_t
N
=
ins
.
size
();
for
(
size_t
i
=
0
;
i
<
N
;
++
i
)
{
VLOG
(
1
)
<<
" inference (X) "
<<
framework
::
product
(
ins
[
i
]
->
dims
())
<<
" (Out)"
<<
framework
::
product
(
outs
[
i
]
->
dims
());
}
for
(
size_t
i
=
0
;
i
<
ins
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
ins
.
size
();
++
i
)
{
VLOG
(
1
)
<<
" invoke allreduce. send "
<<
ins
[
i
]
->
numel
()
<<
" recv "
<<
outs
[
i
]
->
numel
();
PADDLE_ENFORCE
(
platform
::
dynload
::
ncclAllReduce
(
PADDLE_ENFORCE
(
platform
::
dynload
::
ncclAllReduce
(
ins
[
i
]
->
data
<
T
>
(),
outs
[
i
]
->
mutable_data
<
T
>
(
ctx
.
GetPlace
()),
ins
[
i
]
->
data
<
T
>
(),
outs
[
i
]
->
mutable_data
<
T
>
(
ctx
.
GetPlace
()),
outs
[
i
]
->
numel
()
*
sizeof
(
T
)
,
NCCLTypeWrapper
<
T
>::
type
,
ncclSum
,
outs
[
i
]
->
numel
(),
NCCLTypeWrapper
<
T
>::
type
,
ncclSum
,
comm
->
comms_
[
idx
],
stream
));
comm
->
comms_
[
idx
],
stream
));
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
stream
));
PADDLE_ENFORCE
(
cudaStreamSynchronize
(
stream
));
VLOG
(
1
)
<<
" finished allreduce. send "
<<
ins
[
i
]
->
numel
()
<<
" recv "
<<
outs
[
i
]
->
numel
();
}
}
}
}
};
};
...
...
paddle/operators/nccl_op.h
已删除
100644 → 0
浏览文件 @
63fb41b3
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#pragma once
#include "paddle/framework/op_registry.h"
#include "paddle/operators/nccl/nccl_gpu_common.h"
#include <string.h>
namespace
paddle
{
namespace
operators
{
using
framework
::
Tensor
;
using
platform
::
Communicator
;
template
<
typename
Type
>
class
NCCLTypeWrapper
;
template
<
>
class
NCCLTypeWrapper
<
float
>
{
public:
static
const
ncclDataType_t
type
=
ncclFloat
;
};
template
<
>
class
NCCLTypeWrapper
<
double
>
{
public:
static
const
ncclDataType_t
type
=
ncclDouble
;
};
template
<
typename
T
>
class
NCCLInitKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
std
::
vector
<
int
>
gpus
=
ctx
.
Attr
<
std
::
vector
<
int
>>
(
"gpus"
);
auto
*
comm
=
ctx
.
Output
<
Communicator
>
(
"Communicator"
);
comm
->
InitAll
(
gpus
);
}
};
}
// namespace operators
}
// namespace paddle
paddle/operators/nccl_op_test.cu
浏览文件 @
5200c657
...
@@ -12,101 +12,211 @@
...
@@ -12,101 +12,211 @@
See the License for the specific language governing permissions and
See the License for the specific language governing permissions and
limitations under the License. */
limitations under the License. */
#define EIGEN_USE_GPU
#include <glog/logging.h>
#include <glog/logging.h>
#include <gtest/gtest.h>
#include <gtest/gtest.h>
#include <thrust/device_vector.h>
#include <thrust/device_vector.h>
#include <memory>
#include <mutex>
#include <thread>
#include <utility>
#include <vector>
#include <vector>
#include "paddle/framework/block_desc.h"
#include "paddle/framework/block_desc.h"
#include "paddle/framework/op_desc.h"
#include "paddle/framework/op_desc.h"
#include "paddle/framework/op_registry.h"
#include "paddle/framework/program_desc.h"
#include "paddle/framework/program_desc.h"
#include "paddle/framework/var_desc.h"
#include "paddle/framework/var_desc.h"
#include "paddle/operators/nccl/nccl_gpu_common.h"
#include "paddle/platform/device_context.h"
#include "paddle/platform/device_context.h"
#include "paddle/platform/enforce.h"
#include "paddle/platform/enforce.h"
#include "paddle/platform/gpu_info.h"
#include "paddle/platform/gpu_info.h"
#include "paddle/platform/place.h"
#include "paddle/platform/place.h"
USE_CPU_ONLY_OP
(
ncclInit
);
#include "paddle/framework/op_registry.h"
USE_NO_KERNEL_OP
(
ncclInit
);
USE_GPU_ONLY_OP
(
ncclAllReduce
);
USE_GPU_ONLY_OP
(
ncclAllReduce
);
USE_GPU_ONLY_OP
(
ncclReduce
);
USE_GPU_ONLY_OP
(
ncclReduce
);
USE_GPU_ONLY_OP
(
ncclBcastSend
);
USE_GPU_ONLY_OP
(
ncclBcastSend
);
USE_GPU_ONLY_OP
(
ncclBcastRecv
);
USE_GPU_ONLY_OP
(
ncclBcastRecv
);
namespace
f
=
paddle
::
framework
;
namespace
p
=
paddle
::
platform
;
static
std
::
vector
<
int
>
gpu_list
;
static
std
::
vector
<
int
>
gpu_list
;
namespace
f
=
paddle
::
framework
;
// ncclInitOp with desc
namespace
ops
=
paddle
::
operators
;
// TEST(NCCL, ncclInitOp) {
// f::ProgramDescBind program;
void
AddOp
(
const
std
::
string
&
type
,
const
f
::
VariableNameMap
&
inputs
,
// f::BlockDescBind *block = program.Block(0);
const
f
::
VariableNameMap
&
outputs
,
f
::
AttributeMap
attrs
,
// f::OpDescBind *op_desc = block->AppendOp();
paddle
::
framework
::
BlockDescBind
*
block
)
{
for
(
auto
kv
:
outputs
)
{
// op_desc->SetType("ncclInit");
for
(
auto
v
:
kv
.
second
)
{
// op_desc->SetOutput("Communicator", {"x1"});
auto
var
=
block
->
Var
(
v
);
// op_desc->SetAttr("gpus", {gpu_list});
var
->
SetDataType
(
paddle
::
framework
::
DataType
::
FP32
);
// f::Scope g_scope;
}
// p::DeviceContext *ctx =
// new p::CPUDeviceContext(p::CPUPlace());
// auto *var = g_scope.Var("x1");
// var->GetMutable<p::Communicator>();
// auto op = f::OpRegistry::CreateOp(*op_desc);
// VLOG(1) << "invoke NCCLInitOp.";
// op->Run(g_scope, *ctx);
// VLOG(1) << "NCCLInitOp finished.";
// }
// test data amount
static
const
f
::
DDim
kDims
=
{
100
,
100
};
static
std
::
vector
<
p
::
DeviceContext
*>
dev_ctxs
;
void
CreateContext
()
{
for
(
size_t
i
=
0
;
i
<
gpu_list
.
size
();
++
i
)
{
p
::
GPUPlace
place
(
i
);
VLOG
(
1
)
<<
"create devicecontext : "
<<
i
;
dev_ctxs
.
emplace_back
(
new
p
::
CUDADeviceContext
(
place
));
}
}
}
auto
op
=
block
->
AppendOp
();
void
DestroyContext
()
{
op
->
SetType
(
type
);
for
(
size_t
i
=
0
;
i
<
gpu_list
.
size
();
++
i
)
{
for
(
auto
&
kv
:
inputs
)
{
delete
dev_ctxs
[
i
];
op
->
SetInput
(
kv
.
first
,
kv
.
second
);
}
}
for
(
auto
&
kv
:
outputs
)
{
op
->
SetOutput
(
kv
.
first
,
kv
.
second
);
}
op
->
SetAttrMap
(
attrs
);
}
}
// ncclInitOp with desc
// global scope
TEST
(
NCCL
,
ncclInitOp
)
{
static
f
::
Scope
g_scope
;
std
::
mutex
mu
;
template
<
class
T
>
void
DeviceProgram
(
int
gpu_id
,
const
f
::
OpDescBind
&
op_desc
,
f
::
Scope
*
scope
)
{
std
::
unique_lock
<
std
::
mutex
>
lk
(
mu
);
f
::
ProgramDescBind
program
;
f
::
ProgramDescBind
program
;
f
::
BlockDescBind
*
block
=
program
.
Block
(
0
);
f
::
BlockDescBind
*
block
=
program
.
Block
(
0
);
f
::
OpDescBind
*
op_desc
=
block
->
AppendOp
();
f
::
OpDescBind
*
op1
=
block
->
AppendOp
();
*
op1
=
op_desc
;
op_desc
->
SetType
(
"ncclInit"
);
op_desc
->
SetOutput
(
"Communicator"
,
{
"x1"
});
p
::
GPUPlace
place
(
gpu_id
);
op_desc
->
SetAttr
(
"gpus"
,
{
gpu_list
});
// p::DeviceContext *ctx =
f
::
Scope
g_scope
;
// new p::CUDADeviceContext(place);
paddle
::
platform
::
DeviceContext
*
ctx
=
p
::
DeviceContext
*
ctx
=
dev_ctxs
.
at
(
gpu_id
);
new
paddle
::
platform
::
CPUDeviceContext
(
paddle
::
platform
::
CPUPlace
());
VLOG
(
1
)
<<
"device context : "
<<
dev_ctxs
.
size
()
<<
" gpu_id "
<<
gpu_id
;
auto
*
var
=
g_scope
.
Var
(
"x1"
);
// f::Scope &local_scope = g_scope.NewScope();
var
->
GetMutable
<
paddle
::
platform
::
Communicator
>
();
auto
*
send_tensor
=
scope
->
Var
(
"st"
)
->
GetMutable
<
f
::
LoDTensor
>
();
auto
op
=
f
::
OpRegistry
::
CreateOp
(
*
op_desc
);
auto
*
recv_tensor
=
scope
->
Var
(
"rt"
)
->
GetMutable
<
f
::
LoDTensor
>
();
VLOG
(
1
)
<<
"invoke NCCLInitOp."
;
send_tensor
->
Resize
(
kDims
);
op
->
Run
(
g_scope
,
*
ctx
);
send_tensor
->
mutable_data
<
T
>
(
kDims
,
place
);
VLOG
(
1
)
<<
"NCCLInitOp finished."
;
// recv_tensor->mutable_data<T>(kDims, place);
std
::
vector
<
T
>
send_vector
(
f
::
product
(
kDims
),
gpu_id
);
send_tensor
->
CopyFromVector
<
T
>
(
send_vector
,
*
ctx
);
lk
.
unlock
();
PADDLE_ENFORCE
(
send_tensor
->
numel
()
==
f
::
product
(
kDims
),
"Tensor numel not match!"
);
ctx
->
Wait
();
VLOG
(
1
)
<<
send_tensor
->
numel
()
<<
" element in send tensor"
;
auto
op
=
f
::
OpRegistry
::
CreateOp
(
*
op1
);
VLOG
(
1
)
<<
"Device : "
<<
gpu_id
<<
" invoke "
<<
op_desc
.
Type
();
op
->
Run
(
*
scope
,
*
ctx
);
VLOG
(
1
)
<<
"Device : "
<<
gpu_id
<<
" finished "
<<
op_desc
.
Type
();
}
}
// ncclAllReduceOp with desc
// ncclAllReduceOp with desc
TEST
(
NCCL
,
nccl
Init
Op
)
{
TEST
(
NCCL
,
nccl
AllReduce
Op
)
{
f
::
ProgramDescBind
program
;
f
::
ProgramDescBind
program
;
f
::
BlockDescBind
*
block
=
program
.
Block
(
0
);
f
::
BlockDescBind
*
block
=
program
.
Block
(
0
);
f
::
OpDescBind
*
op_desc
=
block
->
AppendOp
();
f
::
OpDescBind
*
op1
=
block
->
AppendOp
();
p
::
DeviceContext
*
ctx
=
new
p
::
CPUDeviceContext
(
p
::
CPUPlace
());
op_desc
->
SetType
(
"ncclAllReduce"
);
CreateContext
(
);
op_desc
->
SetOutput
(
"Communicator"
,
{
"x1"
});
op1
->
SetType
(
"ncclInit"
);
op_desc
->
SetAttr
(
"gpus"
,
{
gpu_list
});
op1
->
SetOutput
(
"Communicator"
,
{
"comm"
});
f
::
Scope
g_scope
;
op1
->
SetAttr
(
"gpus"
,
{
gpu_list
});
paddle
::
platform
::
DeviceContext
*
ctx
=
new
paddle
::
platform
::
CPUDeviceContext
(
paddle
::
platform
::
CPUPlace
());
auto
*
var
=
g_scope
.
Var
(
"
x1
"
);
auto
*
var
=
g_scope
.
Var
(
"
comm
"
);
var
->
GetMutable
<
p
addle
::
platform
::
Communicator
>
();
var
->
GetMutable
<
p
::
Communicator
>
();
auto
op
=
f
::
OpRegistry
::
CreateOp
(
*
op
_desc
);
auto
op
=
f
::
OpRegistry
::
CreateOp
(
*
op
1
);
VLOG
(
1
)
<<
"invoke NCCLInitOp."
;
VLOG
(
1
)
<<
"invoke NCCLInitOp."
;
op
->
Run
(
g_scope
,
*
ctx
);
op
->
Run
(
g_scope
,
*
ctx
);
VLOG
(
1
)
<<
"NCCLInitOp finished."
;
VLOG
(
1
)
<<
"NCCLInitOp finished."
;
delete
ctx
;
f
::
OpDescBind
*
op2
=
new
f
::
OpDescBind
;
op2
->
SetType
(
"ncclAllReduce"
);
op2
->
SetInput
(
"X"
,
{
"st"
});
op2
->
SetInput
(
"Communicator"
,
{
"comm"
});
op2
->
SetOutput
(
"Out"
,
{
"rt"
});
std
::
vector
<
std
::
thread
>
ths
;
for
(
size_t
i
=
0
;
i
<
gpu_list
.
size
();
++
i
)
{
std
::
thread
th
(
DeviceProgram
<
float
>
,
gpu_list
[
i
],
*
op2
,
&
g_scope
.
NewScope
());
// std::thread th([=](){
// VLOG(1) << "thread id created : " << i;
// return 1;});
ths
.
emplace_back
(
std
::
move
(
th
));
}
for
(
size_t
i
=
0
;
i
<
gpu_list
.
size
();
++
i
)
{
VLOG
(
1
)
<<
" thread joined! "
<<
i
;
ths
[
i
].
join
();
}
VLOG
(
1
)
<<
" main thread joined!"
;
delete
op2
;
g_scope
.
~
Scope
();
DestroyContext
();
VLOG
(
1
)
<<
" destory contexts"
;
}
}
// ncclBcastOp with desc
// TEST(NCCL, ncclBcastOp) {
// f::ProgramDescBind program;
// f::BlockDescBind *block = program.Block(0);
// f::OpDescBind *op1= block->AppendOp();
// p::DeviceContext *ctx =
// new p::CPUDeviceContext(p::CPUPlace());
// op1->SetType("ncclInit");
// op1->SetOutput("Communicator", {"comm"});
// op1->SetAttr("gpus", {gpu_list});
// auto *var = g_scope.Var("comm");
// var->GetMutable<p::Communicator>();
// auto op = f::OpRegistry::CreateOp(*op1);
// VLOG(1) << "invoke NCCLInitOp.";
// op->Run(g_scope, *ctx);
// VLOG(1) << "NCCLInitOp finished.";
// f::OpDescBind *op2 = new f::OpDescBind;
// op2->SetType("ncclBcastSend");
// op2->SetInput("X", {"st"});
// op2->SetInput("Communicator", {"comm"});
// op2->SetOutput("Out", {"rt"});
// std::vector<std::thread> ths;
// for (size_t i=0; i < gpu_list.size(); ++i) {
// std::thread th(DeviceProgram<float>, gpu_list[i], *op2);
// ths.emplace_back(std::move(th));
// }
// for (size_t i=0; i < gpu_list.size(); ++i) {
// ths[i].join();
// }
// }
int
main
(
int
argc
,
char
**
argv
)
{
int
main
(
int
argc
,
char
**
argv
)
{
static
int
dev_count
=
paddle
::
platform
::
GetCUDADeviceCount
();
const
int
dev_count
=
p
::
GetCUDADeviceCount
();
if
(
dev_count
<=
1
)
{
if
(
dev_count
<=
1
)
{
LOG
(
WARNING
)
LOG
(
WARNING
)
<<
"Cannot test multi-gpu nccl, because the CUDA device count is "
<<
"Cannot test multi-gpu nccl, because the CUDA device count is "
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录