Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
0f26cee1
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
0f26cee1
编写于
8月 27, 2020
作者:
S
seiriosPlus
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
add large scale optimizer fuse
上级
61284c0a
变更
1
显示空白变更内容
内联
并排
Showing
1 changed file
with
78 addition
and
6 deletion
+78
-6
python/paddle/fluid/incubate/fleet/parameter_server/ir/pserver_pass.py
.../fluid/incubate/fleet/parameter_server/ir/pserver_pass.py
+78
-6
未找到文件。
python/paddle/fluid/incubate/fleet/parameter_server/ir/pserver_pass.py
浏览文件 @
0f26cee1
...
...
@@ -624,6 +624,7 @@ def large_scale_sparse_pass(program, main_program, config, is_startup=False):
value_dims
=
[]
grad
=
None
opt_idx
=
-
1
fuse
=
False
for
op
in
block
.
ops
:
opt_idx
+=
1
...
...
@@ -631,6 +632,9 @@ def large_scale_sparse_pass(program, main_program, config, is_startup=False):
if
op
.
type
not
in
opt_value_map
.
keys
():
continue
if
op
.
type
in
[
"sgd"
,
"adam"
]:
fuse
=
True
grad
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Grad"
)[
0
]]
for
value
in
opt_value_map
[
op
.
type
]:
...
...
@@ -644,7 +648,68 @@ def large_scale_sparse_pass(program, main_program, config, is_startup=False):
if
value_names
:
break
return
grad
,
opt_idx
,
value_names
,
value_dims
,
acture_names
return
grad
,
opt_idx
,
value_names
,
value_dims
,
acture_names
,
fuse
def
add_fuse_large_scale_op
(
block
,
global_block
,
table_name
,
value_names
,
acture_names
,
grad
,
is_entry
,
opt_idx
):
op
=
block
.
ops
[
opt_idx
]
if
op
.
type
==
"sgd"
:
grad
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Grad"
)[
0
]]
# remove origin optimzier op
block
.
_remove_op
(
opt_idx
)
block
.
_insert_op
(
opt_idx
,
type
=
"lookup_sparse_table_fuse_sgd"
,
inputs
=
{
"Grad"
:
grad
},
attrs
=
{
"is_entry"
:
is_entry
,
"tablename"
:
table_name
,
"value_names"
:
value_names
})
elif
op
.
type
==
"adam"
:
grad
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Grad"
)[
0
]]
beta1_pow
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Beta1Pow"
)[
0
]]
beta2_pow
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Beta2Pow"
)[
0
]]
beta1_pow_o
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Beta1PowOut"
)[
0
]]
beta2_pow_o
=
main_program
.
global_block
().
vars
[
op
.
input
(
"Beta2PowOut"
)[
0
]]
beta1
=
op
.
attr
(
'shape'
)
beta2
=
op
.
attr
(
'beta2'
)
epsilon
=
op
.
attr
(
'epsilon'
)
# remove origin optimzier op
block
.
_remove_op
(
opt_idx
)
block
.
_insert_op
(
opt_idx
,
type
=
"lookup_sparse_table_fuse_adam"
,
inputs
=
{
"Grad"
:
grad
,
"Beta1Pow"
:
beta1_pow
,
"Beta2Pow"
:
beta2_pow
},
outputs
=
{
"Beta1PowOut"
:
beta1_pow_o
,
"Beta2PowOut"
:
beta2_pow_o
},
attrs
=
{
"beta1"
:
beta1
,
"beta2"
:
beta2
,
"epsilon"
:
epsilon
,
"is_entry"
:
is_entry
,
"tablename"
:
table_name
,
"value_names"
:
value_names
})
else
:
raise
ValueError
(
"only support sgd/adam optimizer now"
)
def
add_large_scale_op
(
block
,
global_block
,
table_name
,
value_names
,
acture_names
,
grad
,
is_entry
,
opt_idx
):
...
...
@@ -711,20 +776,27 @@ def large_scale_sparse_pass(program, main_program, config, is_startup=False):
for
param
,
blockid
in
param_blockid_map
.
items
():
opt_block
=
program
.
block
(
blockid
)
grad
,
opt_idx
,
value_names
,
value_dims
,
acture_names
=
\
grad
,
opt_idx
,
value_names
,
value_dims
,
acture_names
,
fuse
=
\
get_optimizer_values
(
opt_block
)
entry_attr
=
get_entry_attr
(
param
)
is_entry
=
False
if
entry_attr
==
"none"
else
True
if
fuse
:
add_fuse_large_scale_op
(
opt_block
,
program
.
global_block
(),
param
,
value_names
,
acture_names
,
grad
,
is_entry
,
opt_idx
)
else
:
add_large_scale_op
(
opt_block
,
program
.
global_block
(),
param
,
value_names
,
acture_names
,
grad
,
is_entry
,
opt_idx
)
else
:
large_scale_kv_metas
=
[]
for
param
,
blockid
in
param_blockid_map
.
items
():
opt_block
=
main_program
.
block
(
blockid
)
grad
,
_
,
value_names
,
value_dims
,
acture_names
=
\
grad
,
_
,
value_names
,
value_dims
,
acture_names
,
fuse
=
\
get_optimizer_values
(
opt_block
)
entry_attr
=
get_entry_attr
(
param
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录