Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
3addd568
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
3addd568
编写于
11月 10, 2022
作者:
H
Haohongxiang
提交者:
GitHub
11月 10, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Dygraph] Support grad division to nranks before reduce in sharding stage2 (#47764)
上级
7964119b
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
21 addition
and
19 deletion
+21
-19
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_optimizer_stage2.py
.../meta_parallel/sharding/group_sharded_optimizer_stage2.py
+9
-6
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_stage2.py
...uted/fleet/meta_parallel/sharding/group_sharded_stage2.py
+12
-3
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api_eager.py
...tests/collective/fleet/dygraph_group_sharded_api_eager.py
+0
-10
未找到文件。
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_optimizer_stage2.py
浏览文件 @
3addd568
...
...
@@ -498,12 +498,7 @@ class GroupShardedOptimizerStage2(Optimizer):
with
device_guard
(
self
.
_rank
,
self
.
offload_device
):
self
.
offload_grads
.
buffer
.
zero_
()
def
step
(
self
):
"""
A wrapper for Optimizer's step function to finish the update operation of the optimizer.
"""
# This method won't be called directly by opt.step()!
# The _redefine_opt_step() in class GroupShardedStage2 will wrap this function.
def
_step
(
self
):
if
self
.
_broadcast_overlap
:
# Clear the pre forward hook in the optimizer step.
for
hook_remove
in
self
.
_forward_pre_hook_remove_helper
:
...
...
@@ -536,6 +531,14 @@ class GroupShardedOptimizerStage2(Optimizer):
# Synchronize all the updated shards in between the ranks
self
.
_broadcast_params
()
def
step
(
self
):
"""
A wrapper for Optimizer's step function to finish the update operation of the optimizer.
"""
# This method won't be called directly by opt.step()!
# The _redefine_opt_step() in class GroupShardedStage2 will wrap this function.
self
.
_step
()
def
minimize
(
self
):
raise
RuntimeError
(
"optimizer.minimize() not support now, please use optimizer.step()"
...
...
python/paddle/distributed/fleet/meta_parallel/sharding/group_sharded_stage2.py
浏览文件 @
3addd568
...
...
@@ -225,13 +225,13 @@ class GroupShardedStage2(nn.Layer):
def
_grad_scale
(
self
):
"""
Before the
gradient accumulation, scale the gradient
.
Before the
optimization, scale the gradients before allreduce of dp_group
.
"""
if
self
.
_dp_group
is
None
or
self
.
_dp_group
.
nranks
<=
1
:
scale_factor
=
self
.
_world_size_scaling
return
else
:
scale_factor
=
1.0
/
(
self
.
_
group
.
nranks
*
self
.
_
dp_group
.
nranks
)
scale_factor
=
1.0
/
(
self
.
_dp_group
.
nranks
)
# Scale grad storages
for
dtype
in
self
.
_grad_storages
.
keys
():
...
...
@@ -366,6 +366,13 @@ class GroupShardedStage2(nn.Layer):
),
"Only support comm overlap strategy for single optimizer"
self
.
_sharding_optimizers
[
0
].
_set_reduce_overlap
(
reduce_overlap
)
def
_get_scaled_grad_fn
(
self
):
@
paddle
.
autograd
.
no_grad
()
def
scale
(
grad
):
grad
.
scale_
(
self
.
_world_size_scaling
)
return
scale
def
_get_reduce_fn
(
self
,
index
,
param
,
dst_rank
):
"""
There are two ways to reduce gradient.
...
...
@@ -510,6 +517,8 @@ class GroupShardedStage2(nn.Layer):
return
for
index
,
param
in
enumerate
(
self
.
_trainable_params
):
param
.
_register_grad_hook
(
self
.
_get_scaled_grad_fn
())
dst_rank
=
self
.
_trainable_param2rank
[
param
.
name
]
reduce_function
=
self
.
_get_reduce_fn
(
index
,
param
,
dst_rank
)
...
...
python/paddle/fluid/tests/unittests/collective/fleet/dygraph_group_sharded_api_eager.py
浏览文件 @
3addd568
...
...
@@ -153,16 +153,6 @@ def test_sharding_api():
list
(
range
(
paddle
.
distributed
.
get_world_size
()))
)
stage2_dp_params
=
train_mlp
(
mlp1
,
shard_level
=
"os_g"
,
use_multi_precision
=
True
,
output_dir
=
output_dir
,
amp_level
=
'O2'
,
sync_buffers
=
True
,
dp_group
=
dp_group
,
)
# fp16
stage2_params
=
train_mlp
(
mlp1
,
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录