Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
507afa8a
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
507afa8a
编写于
10月 17, 2019
作者:
W
WangXi
提交者:
gongweibao
10月 18, 2019
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Fix dgc nan by stripping nccl from sparseReduce. (#20630)
上级
c1710e91
变更
11
隐藏空白更改
内联
并排
Showing
11 changed file
with
121 addition
and
40 deletion
+121
-40
cmake/external/dgc.cmake
cmake/external/dgc.cmake
+2
-2
paddle/fluid/framework/details/all_reduce_op_handle.cc
paddle/fluid/framework/details/all_reduce_op_handle.cc
+4
-0
paddle/fluid/framework/details/all_reduce_op_handle.h
paddle/fluid/framework/details/all_reduce_op_handle.h
+2
-0
paddle/fluid/framework/details/sparse_all_reduce_op_handle.cc
...le/fluid/framework/details/sparse_all_reduce_op_handle.cc
+38
-6
paddle/fluid/framework/details/sparse_all_reduce_op_handle.h
paddle/fluid/framework/details/sparse_all_reduce_op_handle.h
+3
-0
paddle/fluid/platform/init.cc
paddle/fluid/platform/init.cc
+0
-17
paddle/fluid/platform/init.h
paddle/fluid/platform/init.h
+0
-2
paddle/fluid/pybind/pybind.cc
paddle/fluid/pybind/pybind.cc
+0
-1
python/paddle/fluid/optimizer.py
python/paddle/fluid/optimizer.py
+0
-2
python/paddle/fluid/tests/unittests/test_dist_base.py
python/paddle/fluid/tests/unittests/test_dist_base.py
+54
-10
python/paddle/fluid/tests/unittests/test_dist_mnist_dgc_nccl.py
.../paddle/fluid/tests/unittests/test_dist_mnist_dgc_nccl.py
+18
-0
未找到文件。
cmake/external/dgc.cmake
浏览文件 @
507afa8a
...
...
@@ -23,8 +23,8 @@ INCLUDE_DIRECTORIES(${DGC_INCLUDE_DIR})
ExternalProject_Add
(
extern_dgc
${
EXTERNAL_PROJECT_LOG_ARGS
}
URL
"http://fleet.bj.bcebos.com/collective.tgz"
URL_MD5
"
015d565156c3de4e30fe25473f47e7a9
"
URL
"http://fleet.bj.bcebos.com/collective
_ef2216a
.tgz"
URL_MD5
"
2f67549fd5f1262383d83289abc4f88f
"
SOURCE_DIR
"
${
DGC_SOURCES_DIR
}
"
CONFIGURE_COMMAND
""
BUILD_COMMAND make -j
...
...
paddle/fluid/framework/details/all_reduce_op_handle.cc
浏览文件 @
507afa8a
...
...
@@ -171,6 +171,10 @@ void AllReduceOpHandle::NCCLAllReduceFunc(
}
});
SyncNCCLAllReduce
();
}
void
AllReduceOpHandle
::
SyncNCCLAllReduce
()
{
if
(
FLAGS_sync_nccl_allreduce
)
{
for
(
auto
&
p
:
places_
)
{
int
dev_id
=
boost
::
get
<
platform
::
CUDAPlace
>
(
p
).
device
;
...
...
paddle/fluid/framework/details/all_reduce_op_handle.h
浏览文件 @
507afa8a
...
...
@@ -63,6 +63,8 @@ class AllReduceOpHandle : public OpHandleBase {
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
void
NCCLAllReduceFunc
(
const
std
::
vector
<
std
::
function
<
void
()
>>
&
all_reduce_calls
);
void
SyncNCCLAllReduce
();
#endif
void
AllReduceImpl
(
const
std
::
vector
<
VarHandle
*>
&
in_var_handles
,
...
...
paddle/fluid/framework/details/sparse_all_reduce_op_handle.cc
浏览文件 @
507afa8a
...
...
@@ -20,6 +20,7 @@
#include "paddle/fluid/framework/details/variable_visitor.h"
#include "paddle/fluid/framework/operator.h"
#include "paddle/fluid/memory/malloc.h"
#include "paddle/fluid/platform/cuda_device_guard.h"
#include "paddle/fluid/platform/gpu_info.h"
#include "paddle/fluid/platform/profiler.h"
...
...
@@ -105,7 +106,8 @@ void SparseAllReduceOpHandle::RunImplEncoded() {
size_t
in_numel
=
0
;
size_t
out_numel
=
0
;
PADDLE_ENFORCE
(
nranks_
>
1
);
std
::
vector
<
std
::
function
<
void
()
>>
all_reduce_calls
;
std
::
vector
<
std
::
function
<
void
()
>>
all_gather_calls
;
std
::
vector
<
std
::
function
<
void
()
>>
sparse_reduce_calls
;
std
::
vector
<
memory
::
AllocationPtr
>
allocations
;
...
...
@@ -141,15 +143,45 @@ void SparseAllReduceOpHandle::RunImplEncoded() {
<<
", nranks:"
<<
nranks_
<<
", gather_buf size:"
<<
buf_size
<<
", k:"
<<
k
<<
", place:"
<<
place
<<
", dtype:"
<<
dtype
;
all_reduce_calls
.
emplace_back
([
=
]
{
PADDLE_ENFORCE
(
paddle
::
communication
::
dgc
::
sparseAllGReduce
(
in_tensor_buf
,
gather_buff
,
k
,
out_tensor_buf
,
out_numel
,
comm
,
stream
));
all_gather_calls
.
emplace_back
([
=
]
{
PADDLE_ENFORCE_CUDA_SUCCESS
(
platform
::
dynload
::
ncclAllGather
(
in_tensor_buf
,
gather_buff
,
2
*
k
,
static_cast
<
ncclDataType_t
>
(
dtype
),
comm
,
stream
));
});
sparse_reduce_calls
.
emplace_back
([
=
]
{
platform
::
CUDADeviceGuard
guard
(
dev_id
);
PADDLE_ENFORCE_EQ
(
paddle
::
communication
::
dgc
::
sparseReduce
(
gather_buff
,
k
,
out_tensor_buf
,
static_cast
<
int
>
(
out_numel
),
nranks_
,
stream
),
true
);
});
}
WaitInputVarGenerated
();
NCCLAllReduceFunc
(
all_reduce_calls
);
SparseAllReduceFunc
(
all_gather_calls
,
sparse_reduce_calls
);
}
void
SparseAllReduceOpHandle
::
SparseAllReduceFunc
(
const
std
::
vector
<
std
::
function
<
void
()
>>
&
all_gather_calls
,
const
std
::
vector
<
std
::
function
<
void
()
>>
&
sparse_reduce_calls
)
{
this
->
RunAndRecordEvent
([
&
]
{
if
(
all_gather_calls
.
size
()
==
1UL
)
{
// Do not use NCCLGroup when manage NCCL by per thread per device
all_gather_calls
[
0
]();
}
else
{
platform
::
NCCLGroupGuard
guard
;
for
(
auto
&
call
:
all_gather_calls
)
{
call
();
}
}
for
(
auto
&
call
:
sparse_reduce_calls
)
{
call
();
}
});
SyncNCCLAllReduce
();
}
int
SparseAllReduceOpHandle
::
GetKValue
(
const
std
::
string
&
grad_name
)
{
...
...
paddle/fluid/framework/details/sparse_all_reduce_op_handle.h
浏览文件 @
507afa8a
...
...
@@ -43,6 +43,9 @@ class SparseAllReduceOpHandle : public AllReduceOpHandle {
int
GetKValue
(
const
std
::
string
&
grad_name
);
bool
IsEncoded
();
void
RunImplEncoded
();
void
SparseAllReduceFunc
(
const
std
::
vector
<
std
::
function
<
void
()
>>
&
all_gather_calls
,
const
std
::
vector
<
std
::
function
<
void
()
>>
&
sparse_reduce_calls
);
private:
bool
is_encoded_
{
false
};
...
...
paddle/fluid/platform/init.cc
浏览文件 @
507afa8a
...
...
@@ -32,9 +32,6 @@ limitations under the License. */
#include "paddle/fluid/platform/init.h"
#include "paddle/fluid/platform/place.h"
#include "paddle/fluid/string/piece.h"
#if defined(PADDLE_WITH_DGC)
#include "dgc/dgc.h"
#endif
DECLARE_int32
(
paddle_num_threads
);
DEFINE_int32
(
multiple_of_cupti_buffer_size
,
1
,
...
...
@@ -51,10 +48,6 @@ namespace framework {
std
::
once_flag
gflags_init_flag
;
std
::
once_flag
p2p_init_flag
;
#if defined(PADDLE_WITH_CUDA) && !defined(_WIN32)
std
::
once_flag
dgc_init_flag
;
#endif
void
InitGflags
(
std
::
vector
<
std
::
string
>
argv
)
{
std
::
call_once
(
gflags_init_flag
,
[
&
]()
{
FLAGS_logtostderr
=
true
;
...
...
@@ -229,15 +222,5 @@ void InitGLOG(const std::string &prog_name) {
#endif
}
#if defined(PADDLE_WITH_DGC)
void
InitDGC
()
{
std
::
call_once
(
dgc_init_flag
,
[]()
{
PADDLE_ENFORCE
(
paddle
::
communication
::
dgc
::
dynloadNcclLib
());
});
}
#else
void
InitDGC
()
{}
#endif
}
// namespace framework
}
// namespace paddle
paddle/fluid/platform/init.h
浏览文件 @
507afa8a
...
...
@@ -30,8 +30,6 @@ void InitDevices(bool init_p2p);
void
InitDevices
(
bool
init_p2p
,
const
std
::
vector
<
int
>
devices
);
void
InitDGC
();
#ifndef _WIN32
void
SignalHandle
(
const
char
*
data
,
int
size
);
#endif
...
...
paddle/fluid/pybind/pybind.cc
浏览文件 @
507afa8a
...
...
@@ -1354,7 +1354,6 @@ All parameter, weight, gradient are variables in Paddle.
m
.
def
(
"init_gflags"
,
framework
::
InitGflags
);
m
.
def
(
"init_glog"
,
framework
::
InitGLOG
);
m
.
def
(
"init_dgc"
,
framework
::
InitDGC
);
m
.
def
(
"load_op_library"
,
framework
::
LoadOpLib
);
m
.
def
(
"init_devices"
,
[](
bool
init_p2p
)
{
framework
::
InitDevices
(
init_p2p
);
});
...
...
python/paddle/fluid/optimizer.py
浏览文件 @
507afa8a
...
...
@@ -959,8 +959,6 @@ class DGCMomentumOptimizer(MomentumOptimizer):
super
(
DGCMomentumOptimizer
,
self
).
__init__
(
learning_rate
,
momentum
,
use_nesterov
,
regularization
,
name
)
core
.
init_dgc
()
def
_add_auto_increment_var
(
self
,
counter_name
,
begin
,
step
=
1
):
helper
=
LayerHelper
(
'global_step_counter'
)
counter
,
is_new_var
=
helper
.
create_or_get_global_variable
(
...
...
python/paddle/fluid/tests/unittests/test_dist_base.py
浏览文件 @
507afa8a
...
...
@@ -580,7 +580,8 @@ class TestDistBase(unittest.TestCase):
check_error_log
=
False
,
batch_size
=
DEFAULT_BATCH_SIZE
,
batch_merge_repeat
=
1
,
log_name
=
""
):
log_name
=
""
,
gpus
=
"0"
):
cmd
=
self
.
_python_interp
...
...
@@ -600,13 +601,17 @@ class TestDistBase(unittest.TestCase):
if
self
.
__use_cuda
:
cmd
+=
" --use_cuda"
env_local
=
{
"CUDA_VISIBLE_DEVICES"
:
"0"
,
"CUDA_VISIBLE_DEVICES"
:
gpus
,
"PADDLE_TRAINERS_NUM"
:
"1"
,
"PADDLE_TRAINER_ID"
:
"0"
}
else
:
env_local
=
{
'CPU_NUM'
:
'1'
}
# not use dgc in single card
if
len
(
gpus
)
>
1
and
self
.
_use_dgc
:
cmd
+=
" --use_dgc"
env_local
.
update
(
envs
)
print
(
"local_cmd: {}, env: {}"
.
format
(
cmd
,
env_local
))
...
...
@@ -833,12 +838,7 @@ class TestDistBase(unittest.TestCase):
print
(
"outs[1]:"
,
outs
[
1
])
return
pickle
.
loads
(
outs
[
0
]),
pickle
.
loads
(
outs
[
1
])
def
check_with_place
(
self
,
model_file
,
delta
=
1e-3
,
check_error_log
=
False
,
need_envs
=
{},
log_name
=
""
):
def
_get_required_envs
(
self
,
check_error_log
=
False
,
need_envs
=
{}):
# TODO(typhoonzero): should auto adapt GPU count on the machine.
required_envs
=
{
"PATH"
:
os
.
getenv
(
"PATH"
,
""
),
...
...
@@ -854,8 +854,6 @@ class TestDistBase(unittest.TestCase):
"NCCL_SHM_DISABLE"
:
"1"
}
required_envs
.
update
(
need_envs
)
if
check_error_log
:
required_envs
[
"GLOG_vmodule"
]
=
\
"fused_all_reduce_op_handle=10,all_reduce_op_handle=10,alloc_continuous_space_op=10,fuse_all_reduce_op_pass=10,"
\
...
...
@@ -863,6 +861,17 @@ class TestDistBase(unittest.TestCase):
"sparse_all_reduce_op_handle=10,gen_nccl_id_op=10"
required_envs
[
"GLOG_logtostderr"
]
=
"1"
required_envs
.
update
(
need_envs
)
return
required_envs
def
check_with_place
(
self
,
model_file
,
delta
=
1e-3
,
check_error_log
=
False
,
need_envs
=
{},
log_name
=
""
):
required_envs
=
self
.
_get_required_envs
(
check_error_log
,
need_envs
)
local_losses
\
=
self
.
_run_local
(
model_file
,
required_envs
,
check_error_log
,
log_name
=
log_name
)
...
...
@@ -893,3 +902,38 @@ class TestDistBase(unittest.TestCase):
dist_loss
=
(
np
.
array
([
tr0_loss
])
+
np
.
array
([
tr1_loss
]))
/
2
print
(
"======="
,
local_loss
,
":"
,
dist_loss
[
0
],
"======="
)
self
.
assertAlmostEqual
(
local_loss
,
dist_loss
[
0
],
delta
=
delta
)
def
check_with_place_multi_cards
(
self
,
model_file
,
delta
=
1e-3
,
check_error_log
=
False
,
need_envs
=
{},
log_name
=
""
):
# need open p2p or shm otherwise multi cards mode will hang
need_envs
.
update
({
"NCCL_P2P_DISABLE"
:
"0"
,
"NCCL_SHM_DISABLE"
:
"0"
})
required_envs
=
self
.
_get_required_envs
(
check_error_log
,
need_envs
)
if
self
.
_use_dgc
:
multi_cards_losses
=
self
.
_run_local
(
model_file
,
required_envs
,
check_error_log
,
log_name
=
log_name
+
"_dgc_2cards"
,
gpus
=
"0,1"
)
self
.
_use_dgc
=
False
base_losses
=
self
.
_run_local
(
model_file
,
required_envs
,
check_error_log
,
log_name
=
log_name
+
"_base_2cards"
,
gpus
=
"0,1"
)
self
.
_use_dgc
=
True
for
step_id
in
range
(
RUN_STEP
):
base_loss
=
base_losses
[
step_id
]
multi_cards_loss
=
multi_cards_losses
[
step_id
]
print
(
"======="
,
base_loss
,
":"
,
multi_cards_loss
,
"======="
)
self
.
assertAlmostEqual
(
base_loss
,
multi_cards_loss
,
delta
=
delta
)
python/paddle/fluid/tests/unittests/test_dist_mnist_dgc_nccl.py
浏览文件 @
507afa8a
...
...
@@ -38,5 +38,23 @@ class TestDistMnistNCCL2DGC(TestDistBase):
log_name
=
flag_name
)
class
TestDistMnistNCCL2DGCMultiCards
(
TestDistBase
):
def
_setup_config
(
self
):
self
.
_sync_mode
=
True
self
.
_use_reduce
=
False
self
.
_use_reader_alloc
=
False
self
.
_nccl2_mode
=
True
self
.
_use_dgc
=
True
def
test_dist_train
(
self
):
import
paddle.fluid
as
fluid
if
fluid
.
core
.
is_compiled_with_cuda
():
self
.
check_with_place_multi_cards
(
"dist_mnist.py"
,
delta
=
1e-5
,
check_error_log
=
True
,
log_name
=
flag_name
)
if
__name__
==
"__main__"
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录