Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
s920243400
PaddleDetection
提交
039d783d
P
PaddleDetection
项目概览
s920243400
/
PaddleDetection
与 Fork 源项目一致
Fork自
PaddlePaddle / PaddleDetection
通知
2
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
039d783d
编写于
3月 18, 2019
作者:
Q
Qiao Longfei
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
change communicator_recv_wait_ms to communicator_max_send_grad_num_before_recv
上级
ea0df4e8
变更
3
显示空白变更内容
内联
并排
Showing
3 changed file
with
20 addition
and
7 deletion
+20
-7
paddle/fluid/operators/distributed/communicator.cc
paddle/fluid/operators/distributed/communicator.cc
+17
-6
paddle/fluid/operators/distributed/communicator.h
paddle/fluid/operators/distributed/communicator.h
+2
-0
python/paddle/fluid/__init__.py
python/paddle/fluid/__init__.py
+1
-1
未找到文件。
paddle/fluid/operators/distributed/communicator.cc
浏览文件 @
039d783d
...
@@ -29,7 +29,8 @@ DEFINE_bool(communicator_independent_recv_thread, true,
...
@@ -29,7 +29,8 @@ DEFINE_bool(communicator_independent_recv_thread, true,
"use an independent to recv vars from parameter server"
);
"use an independent to recv vars from parameter server"
);
DEFINE_int32
(
communicator_send_queue_size
,
20
,
DEFINE_int32
(
communicator_send_queue_size
,
20
,
"queue size to recv gradient before send"
);
"queue size to recv gradient before send"
);
DEFINE_int32
(
communicator_recv_wait_ms
,
200
,
"wait time between each recv"
);
DEFINE_int32
(
communicator_max_send_grad_num_before_recv
,
20
,
"max grad num to send before recv parameters"
);
DEFINE_int32
(
communicator_thread_pool_size
,
5
,
"thread num to do send or recv"
);
DEFINE_int32
(
communicator_thread_pool_size
,
5
,
"thread num to do send or recv"
);
DEFINE_int32
(
communicator_max_merge_var_num
,
20
,
DEFINE_int32
(
communicator_max_merge_var_num
,
20
,
"max var num to merge and send"
);
"max var num to merge and send"
);
...
@@ -60,7 +61,8 @@ Communicator::Communicator(const RpcCtxMap &send_varname_to_ctx,
...
@@ -60,7 +61,8 @@ Communicator::Communicator(const RpcCtxMap &send_varname_to_ctx,
<<
FLAGS_communicator_independent_recv_thread
;
<<
FLAGS_communicator_independent_recv_thread
;
VLOG
(
0
)
<<
"communicator_send_queue_size: "
VLOG
(
0
)
<<
"communicator_send_queue_size: "
<<
FLAGS_communicator_send_queue_size
;
<<
FLAGS_communicator_send_queue_size
;
VLOG
(
0
)
<<
"communicator_recv_wait_ms: "
<<
FLAGS_communicator_recv_wait_ms
;
VLOG
(
0
)
<<
"communicator_max_send_grad_num_before_recv: "
<<
FLAGS_communicator_max_send_grad_num_before_recv
;
VLOG
(
0
)
<<
"communicator_thread_pool_size: "
VLOG
(
0
)
<<
"communicator_thread_pool_size: "
<<
FLAGS_communicator_thread_pool_size
;
<<
FLAGS_communicator_thread_pool_size
;
VLOG
(
0
)
<<
"communicator_max_merge_var_num: "
VLOG
(
0
)
<<
"communicator_max_merge_var_num: "
...
@@ -102,6 +104,10 @@ void Communicator::SendThread() {
...
@@ -102,6 +104,10 @@ void Communicator::SendThread() {
while
(
var_queue
->
Size
()
>
0
&&
while
(
var_queue
->
Size
()
>
0
&&
merged_var_num
<
FLAGS_communicator_max_merge_var_num
)
{
merged_var_num
<
FLAGS_communicator_max_merge_var_num
)
{
vars
.
push_back
(
var_queue
->
Pop
());
vars
.
push_back
(
var_queue
->
Pop
());
// only count the send number of the first var
if
(
var_name
==
send_varname_to_queue_
.
begin
()
->
first
)
{
grad_num_
.
fetch_add
(
1
,
std
::
memory_order_relaxed
);
}
merged_var_num
++
;
merged_var_num
++
;
}
}
auto
before_merge
=
GetCurrentUS
();
auto
before_merge
=
GetCurrentUS
();
...
@@ -129,7 +135,7 @@ void Communicator::SendThread() {
...
@@ -129,7 +135,7 @@ void Communicator::SendThread() {
}
}
auto
after_run_send_graph
=
GetCurrentUS
();
auto
after_run_send_graph
=
GetCurrentUS
();
auto
send_graph_use_time
=
after_run_send_graph
-
before_run_send_graph
;
auto
send_graph_use_time
=
after_run_send_graph
-
before_run_send_graph
;
if
(
send_graph_use_time
>
10
)
{
if
(
send_graph_use_time
>
10
0
)
{
VLOG
(
1
)
<<
"run send graph use time "
VLOG
(
1
)
<<
"run send graph use time "
<<
after_run_send_graph
-
before_run_send_graph
;
<<
after_run_send_graph
-
before_run_send_graph
;
}
}
...
@@ -165,9 +171,14 @@ void Communicator::RecvAll() {
...
@@ -165,9 +171,14 @@ void Communicator::RecvAll() {
void
Communicator
::
RecvThread
()
{
void
Communicator
::
RecvThread
()
{
VLOG
(
3
)
<<
"RecvThread start!"
;
VLOG
(
3
)
<<
"RecvThread start!"
;
while
(
running_
)
{
while
(
running_
)
{
auto
grad_num
=
grad_num_
.
load
();
if
(
grad_num
>
FLAGS_communicator_max_send_grad_num_before_recv
)
{
VLOG
(
1
)
<<
"current grad num "
<<
grad_num
;
RecvAll
();
RecvAll
();
std
::
this_thread
::
sleep_for
(
grad_num_
.
store
(
0
);
std
::
chrono
::
milliseconds
(
FLAGS_communicator_recv_wait_ms
));
}
else
{
std
::
this_thread
::
sleep_for
(
std
::
chrono
::
milliseconds
(
10
));
}
}
}
}
}
...
...
paddle/fluid/operators/distributed/communicator.h
浏览文件 @
039d783d
...
@@ -14,6 +14,7 @@ limitations under the License. */
...
@@ -14,6 +14,7 @@ limitations under the License. */
#pragma once
#pragma once
#include <atomic>
#include <deque>
#include <deque>
#include <memory>
#include <memory>
#include <string>
#include <string>
...
@@ -184,6 +185,7 @@ class Communicator {
...
@@ -184,6 +185,7 @@ class Communicator {
std
::
unique_ptr
<
Scope
>
send_scope_
;
// an independent scope
std
::
unique_ptr
<
Scope
>
send_scope_
;
// an independent scope
std
::
unique_ptr
<::
ThreadPool
>
send_threadpool_
{
nullptr
};
std
::
unique_ptr
<::
ThreadPool
>
send_threadpool_
{
nullptr
};
std
::
unique_ptr
<::
ThreadPool
>
recv_threadpool_
{
nullptr
};
std
::
unique_ptr
<::
ThreadPool
>
recv_threadpool_
{
nullptr
};
std
::
atomic_uint
grad_num_
{
0
};
// the num of gradient sent since last recv
// the following code is for initialize the commnunicator
// the following code is for initialize the commnunicator
public:
public:
...
...
python/paddle/fluid/__init__.py
浏览文件 @
039d783d
...
@@ -155,7 +155,7 @@ def __bootstrap__():
...
@@ -155,7 +155,7 @@ def __bootstrap__():
# env for communicator
# env for communicator
read_env_flags
.
append
(
'communicator_independent_recv_thread'
)
read_env_flags
.
append
(
'communicator_independent_recv_thread'
)
read_env_flags
.
append
(
'communicator_send_queue_size'
)
read_env_flags
.
append
(
'communicator_send_queue_size'
)
read_env_flags
.
append
(
'communicator_
recv_wait_ms
'
)
read_env_flags
.
append
(
'communicator_
max_send_grad_num_before_recv
'
)
read_env_flags
.
append
(
'communicator_thread_pool_size'
)
read_env_flags
.
append
(
'communicator_thread_pool_size'
)
read_env_flags
.
append
(
'communicator_max_merge_var_num'
)
read_env_flags
.
append
(
'communicator_max_merge_var_num'
)
read_env_flags
.
append
(
'communicator_fake_rpc'
)
read_env_flags
.
append
(
'communicator_fake_rpc'
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录