Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
MindSpore
book
提交
69fedea0
B
book
项目概览
MindSpore
/
book
通知
3
Star
1
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
B
book
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
69fedea0
编写于
6月 20, 2020
作者:
M
mindspore-ci-bot
提交者:
Gitee
6月 20, 2020
浏览文件
操作
浏览文件
下载
差异文件
!8 Modify the code in chapter13
Merge pull request !8 from ougongchang/master
上级
b6bd481a
7ba26043
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
85 addition
and
68 deletion
+85
-68
chapter13/lenet.py
chapter13/lenet.py
+76
-0
chapter13/lenet_with_summary.py
chapter13/lenet_with_summary.py
+9
-68
未找到文件。
chapter13/lenet.py
0 → 100644
浏览文件 @
69fedea0
# Copyright 2020 Huawei Technologies Co., Ltd
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
# ============================================================================
"""LeNet."""
import
mindspore.nn
as
nn
from
mindspore.common.initializer
import
TruncatedNormal
def
conv
(
in_channels
,
out_channels
,
kernel_size
,
stride
=
1
,
padding
=
0
):
"""weight initial for conv layer"""
weight
=
weight_variable
()
return
nn
.
Conv2d
(
in_channels
,
out_channels
,
kernel_size
=
kernel_size
,
stride
=
stride
,
padding
=
padding
,
weight_init
=
weight
,
has_bias
=
False
,
pad_mode
=
"valid"
)
def
fc_with_initialize
(
input_channels
,
out_channels
):
"""weight initial for fc layer"""
weight
=
weight_variable
()
bias
=
weight_variable
()
return
nn
.
Dense
(
input_channels
,
out_channels
,
weight
,
bias
)
def
weight_variable
():
"""weight initial"""
return
TruncatedNormal
(
0.02
)
class
LeNet5
(
nn
.
Cell
):
"""
Lenet network
Args:
num_class (int): Num classes. Default: 10.
Returns:
Tensor, output tensor
Examples:
>>> LeNet(num_class=10)
"""
def
__init__
(
self
,
num_class
=
10
):
super
(
LeNet5
,
self
).
__init__
()
self
.
num_class
=
num_class
self
.
batch_size
=
32
self
.
conv1
=
conv
(
1
,
6
,
5
)
self
.
conv2
=
conv
(
6
,
16
,
5
)
self
.
fc1
=
fc_with_initialize
(
16
*
5
*
5
,
120
)
self
.
fc2
=
fc_with_initialize
(
120
,
84
)
self
.
fc3
=
fc_with_initialize
(
84
,
self
.
num_class
)
self
.
relu
=
nn
.
ReLU
()
self
.
max_pool2d
=
nn
.
MaxPool2d
(
kernel_size
=
2
,
stride
=
2
)
self
.
flatten
=
nn
.
Flatten
()
def
construct
(
self
,
x
):
x
=
self
.
conv1
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
max_pool2d
(
x
)
x
=
self
.
conv2
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
max_pool2d
(
x
)
x
=
self
.
flatten
(
x
)
x
=
self
.
fc1
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
fc2
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
fc3
(
x
)
return
x
chapter13/lenet_with_summary.py
浏览文件 @
69fedea0
...
...
@@ -21,20 +21,18 @@ from mindspore import nn
from
mindspore
import
context
,
Tensor
from
mindspore.train
import
Model
from
mindspore.ops
import
operations
as
P
from
mindspore.common.initializer
import
TruncatedNormal
from
mindspore.dataset.transforms.vision
import
c_transforms
as
transforms
from
mindspore.dataset.transforms.vision
import
Inter
from
mindspore.dataset.transforms
import
c_transforms
as
C
from
mindspore.ops
import
functional
as
F
from
mindspore.common
import
dtype
as
mstype
from
mindspore.train.callback
import
SummaryStep
from
mindspore.train.summary.summary_record
import
SummaryRecord
from
mindspore.train.callback
import
SummaryCollector
from
.lenet
import
LeNet5
class
CrossEntropyLoss
(
nn
.
Cell
):
"""
Define loss for network
"""
"""Define loss function for network."""
def
__init__
(
self
):
super
(
CrossEntropyLoss
,
self
).
__init__
()
self
.
sm_scalar
=
P
.
ScalarSummary
()
...
...
@@ -52,11 +50,8 @@ class CrossEntropyLoss(nn.Cell):
return
loss
def
create_dataset
(
data_path
,
batch_size
=
32
,
repeat_size
=
1
,
num_parallel_workers
=
1
):
"""
create dataset for train or test
"""
def
create_dataset
(
data_path
,
batch_size
=
32
,
repeat_size
=
1
,
num_parallel_workers
=
1
):
"""Create dataset for train or test."""
# define dataset
mnist_ds
=
dataset
.
MnistDataset
(
data_path
)
...
...
@@ -85,58 +80,6 @@ def create_dataset(data_path, batch_size=32, repeat_size=1,
return
mnist_ds
def
conv
(
in_channels
,
out_channels
,
kernel_size
,
stride
=
1
,
padding
=
0
):
weight
=
weight_variable
()
return
nn
.
Conv2d
(
in_channels
,
out_channels
,
kernel_size
=
kernel_size
,
stride
=
stride
,
padding
=
padding
,
weight_init
=
weight
,
has_bias
=
False
,
pad_mode
=
"valid"
)
def
fc_with_initialize
(
input_channels
,
out_channels
):
weight
=
weight_variable
()
bias
=
weight_variable
()
return
nn
.
Dense
(
input_channels
,
out_channels
,
weight
,
bias
)
def
weight_variable
():
return
TruncatedNormal
(
0.02
)
class
LeNet5
(
nn
.
Cell
):
"""
Lenet network
"""
def
__init__
(
self
):
super
(
LeNet5
,
self
).
__init__
()
self
.
sm_image
=
P
.
ImageSummary
()
self
.
batch_size
=
32
self
.
conv1
=
conv
(
1
,
6
,
5
)
self
.
conv2
=
conv
(
6
,
16
,
5
)
self
.
fc1
=
fc_with_initialize
(
16
*
5
*
5
,
120
)
self
.
fc2
=
fc_with_initialize
(
120
,
84
)
self
.
fc3
=
fc_with_initialize
(
84
,
10
)
self
.
relu
=
nn
.
ReLU
()
self
.
max_pool2d
=
nn
.
MaxPool2d
(
kernel_size
=
2
,
stride
=
2
)
self
.
reshape
=
P
.
Reshape
()
def
construct
(
self
,
x
):
self
.
sm_image
(
"image"
,
x
)
x
=
self
.
conv1
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
max_pool2d
(
x
)
x
=
self
.
conv2
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
max_pool2d
(
x
)
x
=
self
.
reshape
(
x
,
(
self
.
batch_size
,
-
1
))
x
=
self
.
fc1
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
fc2
(
x
)
x
=
self
.
relu
(
x
)
x
=
self
.
fc3
(
x
)
return
x
def
main
(
data_path
,
device_target
=
'Ascend'
,
summary_dir
=
'./summary_dir'
,
learning_rate
=
0.01
):
context
.
set_context
(
mode
=
context
.
GRAPH_MODE
,
device_target
=
device_target
)
...
...
@@ -150,15 +93,13 @@ def main(data_path, device_target='Ascend', summary_dir='./summary_dir', learnin
net_opt
=
nn
.
Momentum
(
network
.
trainable_params
(),
learning_rate
,
momentum
)
model
=
Model
(
network
,
net_loss
,
net_opt
)
# add summary writer
summary_writer
=
SummaryRecord
(
log_dir
=
summary_dir
,
network
=
network
)
summary_callback
=
SummaryStep
(
summary_writer
,
flush_step
=
10
)
# Init SummaryCollector callback to record summary data in model.train or model.eval
summary_collector
=
SummaryCollector
(
summary_dir
=
summary_dir
,
collect_freq
=
10
)
ds
=
create_dataset
(
os
.
path
.
join
(
data_path
,
"train"
),
batch_size
=
batch_size
)
print
(
"============== Starting Training =============="
)
model
.
train
(
epoch_size
,
ds
,
callbacks
=
[
summary_callback
],
dataset_sink_mode
=
False
)
summary_writer
.
close
()
model
.
train
(
epoch_size
,
ds
,
callbacks
=
[
summary_collector
],
dataset_sink_mode
=
False
)
print
(
"============== Train End ====================="
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录