Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
MindSpore
book
提交
260d308a
B
book
项目概览
MindSpore
/
book
通知
3
Star
1
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
B
book
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
260d308a
编写于
9月 04, 2020
作者:
Y
yao_yf
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
adapt parallel interface change
上级
bacd6196
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
7 addition
and
6 deletion
+7
-6
chapter05/resnet/resnet_cifar.py
chapter05/resnet/resnet_cifar.py
+3
-2
chapter07/run_pretrain.py
chapter07/run_pretrain.py
+2
-2
chapter07/src/bert_for_pre_training.py
chapter07/src/bert_for_pre_training.py
+2
-2
未找到文件。
chapter05/resnet/resnet_cifar.py
浏览文件 @
260d308a
...
...
@@ -31,7 +31,8 @@ import mindspore.dataset.transforms.c_transforms as C2
from
mindspore
import
Tensor
from
mindspore.ops
import
operations
as
P
from
mindspore.nn.optim.momentum
import
Momentum
from
mindspore.train.model
import
Model
,
ParallelMode
from
mindspore.train.model
import
Model
from
mindspore.context
import
ParallelMode
from
mindspore
import
context
from
mindspore.train.callback
import
ModelCheckpoint
,
CheckpointConfig
,
LossMonitor
from
mindspore.train.serialization
import
load_checkpoint
,
load_param_into_net
...
...
@@ -113,7 +114,7 @@ def create_dataset(repeat_num=1, training=True):
if
__name__
==
'__main__'
:
if
args_opt
.
mode
==
'train'
and
args_opt
.
run_distribute
:
context
.
set_auto_parallel_context
(
device_num
=
args_opt
.
device_num
,
parallel_mode
=
ParallelMode
.
DATA_PARALLEL
,
mirror
_mean
=
True
)
context
.
set_auto_parallel_context
(
device_num
=
args_opt
.
device_num
,
parallel_mode
=
ParallelMode
.
DATA_PARALLEL
,
gradients
_mean
=
True
)
auto_parallel_context
().
set_all_reduce_fusion_split_indices
([
140
])
init
()
...
...
chapter07/run_pretrain.py
浏览文件 @
260d308a
...
...
@@ -24,7 +24,7 @@ import mindspore.communication.management as D
import
mindspore.common.dtype
as
mstype
from
mindspore
import
context
from
mindspore.train.model
import
Model
from
mindspore.
train.parallel_utils
import
ParallelMode
from
mindspore.
context
import
ParallelMode
from
mindspore.nn.wrap.loss_scale
import
DynamicLossScaleUpdateCell
from
mindspore.train.callback
import
ModelCheckpoint
,
CheckpointConfig
,
TimeMonitor
from
mindspore.train.serialization
import
load_checkpoint
,
load_param_into_net
...
...
@@ -77,7 +77,7 @@ def run_pretrain():
ckpt_save_dir
=
args_opt
.
save_checkpoint_path
+
'ckpt_'
+
str
(
rank
)
+
'/'
context
.
reset_auto_parallel_context
()
context
.
set_auto_parallel_context
(
parallel_mode
=
ParallelMode
.
DATA_PARALLEL
,
mirror
_mean
=
True
,
context
.
set_auto_parallel_context
(
parallel_mode
=
ParallelMode
.
DATA_PARALLEL
,
gradients
_mean
=
True
,
device_num
=
device_num
)
from
mindspore.parallel._auto_parallel_context
import
auto_parallel_context
if
bert_net_cfg
.
num_hidden_layers
==
12
:
...
...
chapter07/src/bert_for_pre_training.py
浏览文件 @
260d308a
...
...
@@ -24,7 +24,7 @@ from mindspore.common.tensor import Tensor
from
mindspore.common.parameter
import
Parameter
,
ParameterTuple
from
mindspore.common
import
dtype
as
mstype
from
mindspore.nn.wrap.grad_reducer
import
DistributedGradReducer
from
mindspore.
train.parallel_utils
import
ParallelMode
from
mindspore.
context
import
ParallelMode
from
mindspore.communication.management
import
get_group_size
from
mindspore
import
context
from
mindspore.ops
import
_selected_ops
...
...
@@ -280,7 +280,7 @@ class BertTrainOneStepCell(nn.Cell):
self
.
reducer_flag
=
True
self
.
grad_reducer
=
None
if
self
.
reducer_flag
:
mean
=
context
.
get_auto_parallel_context
(
"
mirror
_mean"
)
mean
=
context
.
get_auto_parallel_context
(
"
gradients
_mean"
)
degree
=
get_group_size
()
self
.
grad_reducer
=
DistributedGradReducer
(
optimizer
.
parameters
,
mean
,
degree
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录