Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleSeg
提交
33cd5c87
P
PaddleSeg
项目概览
PaddlePaddle
/
PaddleSeg
通知
285
Star
8
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
53
列表
看板
标记
里程碑
合并请求
3
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleSeg
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
53
Issue
53
列表
看板
标记
里程碑
合并请求
3
合并请求
3
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
33cd5c87
编写于
8月 17, 2020
作者:
C
chenguowei01
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
update benchmark
上级
5f72c538
变更
2
显示空白变更内容
内联
并排
Showing
2 changed file
with
56 addition
and
58 deletion
+56
-58
dygraph/benchmark/deeplabv3p.py
dygraph/benchmark/deeplabv3p.py
+27
-31
dygraph/benchmark/hrnet.py
dygraph/benchmark/hrnet.py
+29
-27
未找到文件。
dygraph/benchmark/deeplabv3p.py
浏览文件 @
33cd5c87
...
@@ -13,22 +13,14 @@
...
@@ -13,22 +13,14 @@
# limitations under the License.
# limitations under the License.
import
argparse
import
argparse
import
os
import
sys
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
from
paddle.fluid.dygraph.parallel
import
ParallelEnv
from
paddle.fluid.dygraph.parallel
import
ParallelEnv
from
paddle.fluid.io
import
DataLoader
from
paddle.incubate.hapi.distributed
import
DistributedBatchSampler
from
datasets
import
OpticDiscSeg
,
Cityscapes
from
datasets
import
DATASETS
import
transforms
as
T
import
transforms
as
T
from
models
import
MODELS
from
models
import
MODELS
import
utils.logging
as
logging
from
utils
import
get_environ_info
from
utils
import
get_environ_info
from
utils
import
load_pretrained_model
from
utils
import
resume
from
utils
import
Timer
,
calculate_eta
from
core
import
train
from
core
import
train
...
@@ -48,10 +40,16 @@ def parse_args():
...
@@ -48,10 +40,16 @@ def parse_args():
parser
.
add_argument
(
parser
.
add_argument
(
'--dataset'
,
'--dataset'
,
dest
=
'dataset'
,
dest
=
'dataset'
,
help
=
help
=
"The dataset you want to train, which is one of {}"
.
format
(
"The dataset you want to train, which is one of ('OpticDiscSeg', 'Cityscapes')"
,
str
(
list
(
DATASETS
.
keys
())))
,
type
=
str
,
type
=
str
,
default
=
'OpticDiscSeg'
)
default
=
'OpticDiscSeg'
)
parser
.
add_argument
(
'--dataset_root'
,
dest
=
'dataset_root'
,
help
=
"dataset root directory"
,
type
=
str
,
default
=
None
)
# params of training
# params of training
parser
.
add_argument
(
parser
.
add_argument
(
...
@@ -135,36 +133,38 @@ def main(args):
...
@@ -135,36 +133,38 @@ def main(args):
if
env_info
[
'place'
]
==
'cuda'
and
fluid
.
is_compiled_with_cuda
()
\
if
env_info
[
'place'
]
==
'cuda'
and
fluid
.
is_compiled_with_cuda
()
\
else
fluid
.
CPUPlace
()
else
fluid
.
CPUPlace
()
if
args
.
dataset
.
lower
()
==
'opticdiscseg'
:
if
args
.
dataset
not
in
DATASETS
:
dataset
=
OpticDiscSeg
raise
Exception
(
'`--dataset` is invalid. it should be one of {}'
.
format
(
elif
args
.
dataset
.
lower
()
==
'cityscapes'
:
str
(
list
(
DATASETS
.
keys
()))))
dataset
=
Cityscapes
dataset
=
DATASETS
[
args
.
dataset
]
else
:
raise
Exception
(
"The --dataset set wrong. It should be one of ('OpticDiscSeg', 'Cityscapes')"
)
with
fluid
.
dygraph
.
guard
(
places
):
with
fluid
.
dygraph
.
guard
(
places
):
# Creat dataset reader
# Creat dataset reader
train_transforms
=
T
.
Compose
([
train_transforms
=
T
.
Compose
([
T
.
RandomHorizontalFlip
(
0.5
),
T
.
ResizeStepScaling
(
0.5
,
2.0
,
0.25
),
T
.
ResizeStepScaling
(
0.5
,
2.0
,
0.25
),
T
.
RandomPaddingCrop
(
args
.
input_size
),
T
.
RandomPaddingCrop
(
args
.
input_size
),
T
.
Random
HorizontalFlip
(),
T
.
Random
Distort
(),
T
.
Normalize
()
T
.
Normalize
()
,
])
])
train_dataset
=
dataset
(
transforms
=
train_transforms
,
mode
=
'train'
)
train_dataset
=
dataset
(
dataset_root
=
args
.
dataset_root
,
transforms
=
train_transforms
,
mode
=
'train'
)
eval_dataset
=
None
eval_dataset
=
None
if
args
.
do_eval
:
if
args
.
do_eval
:
eval_transforms
=
T
.
Compose
(
eval_transforms
=
T
.
Compose
(
[
T
.
Padding
((
2049
,
1025
)),
[
T
.
Padding
((
2049
,
1025
)),
T
.
Normalize
()]
T
.
Normalize
()])
)
eval_dataset
=
dataset
(
eval_dataset
=
dataset
(
transforms
=
eval_transforms
,
mode
=
'eval'
)
dataset_root
=
args
.
dataset_root
,
transforms
=
eval_transforms
,
mode
=
'val'
)
if
args
.
model_name
not
in
MODELS
:
if
args
.
model_name
not
in
MODELS
:
raise
Exception
(
raise
Exception
(
'
--model_name
is invalid. it should be one of {}'
.
format
(
'
`--model_name`
is invalid. it should be one of {}'
.
format
(
str
(
list
(
MODELS
.
keys
()))))
str
(
list
(
MODELS
.
keys
()))))
model
=
MODELS
[
args
.
model_name
](
num_classes
=
train_dataset
.
num_classes
)
model
=
MODELS
[
args
.
model_name
](
num_classes
=
train_dataset
.
num_classes
)
...
@@ -174,17 +174,13 @@ def main(args):
...
@@ -174,17 +174,13 @@ def main(args):
args
.
batch_size
*
ParallelEnv
().
nranks
)
args
.
batch_size
*
ParallelEnv
().
nranks
)
decay_step
=
args
.
num_epochs
*
num_steps_each_epoch
decay_step
=
args
.
num_epochs
*
num_steps_each_epoch
lr_decay
=
fluid
.
layers
.
polynomial_decay
(
lr_decay
=
fluid
.
layers
.
polynomial_decay
(
args
.
learning_rate
,
decay_step
,
end_learning_rate
=
0.00001
,
power
=
0.9
)
args
.
learning_rate
,
decay_step
,
end_learning_rate
=
0
,
power
=
0.9
)
optimizer
=
fluid
.
optimizer
.
Momentum
(
optimizer
=
fluid
.
optimizer
.
Momentum
(
lr_decay
,
lr_decay
,
momentum
=
0.9
,
momentum
=
0.9
,
parameter_list
=
model
.
parameters
(),
parameter_list
=
model
.
parameters
(),
#parameter_list=filter(lambda p: p.trainable, model.parameters()),
regularization
=
fluid
.
regularizer
.
L2Decay
(
regularization_coeff
=
4e-5
))
regularization
=
fluid
.
regularizer
.
L2Decay
(
regularization_coeff
=
4e-5
))
train
(
train
(
model
,
model
,
train_dataset
,
train_dataset
,
...
...
dygraph/benchmark/hrnet
_w18_benchmark
.py
→
dygraph/benchmark/hrnet.py
浏览文件 @
33cd5c87
...
@@ -13,22 +13,15 @@
...
@@ -13,22 +13,15 @@
# limitations under the License.
# limitations under the License.
import
argparse
import
argparse
import
os
import
paddle.fluid
as
fluid
import
paddle.fluid
as
fluid
from
paddle.fluid.dygraph.parallel
import
ParallelEnv
from
paddle.fluid.dygraph.parallel
import
ParallelEnv
from
paddle.fluid.io
import
DataLoader
from
paddle.incubate.hapi.distributed
import
DistributedBatchSampler
from
datasets
import
OpticDiscSeg
,
Cityscapes
from
datasets
import
DATASETS
import
transforms
as
T
import
transforms
as
T
from
models
import
MODELS
from
models
import
MODELS
import
utils.logging
as
logging
from
utils
import
get_environ_info
from
utils
import
get_environ_info
from
utils
import
load_pretrained_model
from
core
import
train
from
utils
import
resume
from
utils
import
Timer
,
calculate_eta
from
core
import
train
,
evaluate
def
parse_args
():
def
parse_args
():
...
@@ -47,10 +40,16 @@ def parse_args():
...
@@ -47,10 +40,16 @@ def parse_args():
parser
.
add_argument
(
parser
.
add_argument
(
'--dataset'
,
'--dataset'
,
dest
=
'dataset'
,
dest
=
'dataset'
,
help
=
help
=
"The dataset you want to train, which is one of {}"
.
format
(
"The dataset you want to train, which is one of ('OpticDiscSeg', 'Cityscapes')"
,
str
(
list
(
DATASETS
.
keys
())))
,
type
=
str
,
type
=
str
,
default
=
'Cityscapes'
)
default
=
'OpticDiscSeg'
)
parser
.
add_argument
(
'--dataset_root'
,
dest
=
'dataset_root'
,
help
=
"dataset root directory"
,
type
=
str
,
default
=
None
)
# params of training
# params of training
parser
.
add_argument
(
parser
.
add_argument
(
...
@@ -58,14 +57,14 @@ def parse_args():
...
@@ -58,14 +57,14 @@ def parse_args():
dest
=
"input_size"
,
dest
=
"input_size"
,
help
=
"The image size for net inputs."
,
help
=
"The image size for net inputs."
,
nargs
=
2
,
nargs
=
2
,
default
=
[
1024
,
512
],
default
=
[
512
,
512
],
type
=
int
)
type
=
int
)
parser
.
add_argument
(
parser
.
add_argument
(
'--num_epochs'
,
'--num_epochs'
,
dest
=
'num_epochs'
,
dest
=
'num_epochs'
,
help
=
'Number epochs for training'
,
help
=
'Number epochs for training'
,
type
=
int
,
type
=
int
,
default
=
5
00
)
default
=
1
00
)
parser
.
add_argument
(
parser
.
add_argument
(
'--batch_size'
,
'--batch_size'
,
dest
=
'batch_size'
,
dest
=
'batch_size'
,
...
@@ -107,7 +106,7 @@ def parse_args():
...
@@ -107,7 +106,7 @@ def parse_args():
dest
=
'num_workers'
,
dest
=
'num_workers'
,
help
=
'Num workers for data loader'
,
help
=
'Num workers for data loader'
,
type
=
int
,
type
=
int
,
default
=
2
)
default
=
0
)
parser
.
add_argument
(
parser
.
add_argument
(
'--do_eval'
,
'--do_eval'
,
dest
=
'do_eval'
,
dest
=
'do_eval'
,
...
@@ -134,14 +133,10 @@ def main(args):
...
@@ -134,14 +133,10 @@ def main(args):
if
env_info
[
'place'
]
==
'cuda'
and
fluid
.
is_compiled_with_cuda
()
\
if
env_info
[
'place'
]
==
'cuda'
and
fluid
.
is_compiled_with_cuda
()
\
else
fluid
.
CPUPlace
()
else
fluid
.
CPUPlace
()
if
args
.
dataset
.
lower
()
==
'opticdiscseg'
:
if
args
.
dataset
not
in
DATASETS
:
dataset
=
OpticDiscSeg
raise
Exception
(
'`--dataset` is invalid. it should be one of {}'
.
format
(
elif
args
.
dataset
.
lower
()
==
'cityscapes'
:
str
(
list
(
DATASETS
.
keys
()))))
dataset
=
Cityscapes
dataset
=
DATASETS
[
args
.
dataset
]
else
:
raise
Exception
(
"The --dataset set wrong. It should be one of ('OpticDiscSeg', 'Cityscapes')"
)
with
fluid
.
dygraph
.
guard
(
places
):
with
fluid
.
dygraph
.
guard
(
places
):
# Creat dataset reader
# Creat dataset reader
...
@@ -152,16 +147,22 @@ def main(args):
...
@@ -152,16 +147,22 @@ def main(args):
T
.
RandomDistort
(),
T
.
RandomDistort
(),
T
.
Normalize
(),
T
.
Normalize
(),
])
])
train_dataset
=
dataset
(
transforms
=
train_transforms
,
mode
=
'train'
)
train_dataset
=
dataset
(
dataset_root
=
args
.
dataset_root
,
transforms
=
train_transforms
,
mode
=
'train'
)
eval_dataset
=
None
eval_dataset
=
None
if
args
.
do_eval
:
if
args
.
do_eval
:
eval_transforms
=
T
.
Compose
([
T
.
Normalize
()])
eval_transforms
=
T
.
Compose
([
T
.
Normalize
()])
eval_dataset
=
dataset
(
transforms
=
eval_transforms
,
mode
=
'eval'
)
eval_dataset
=
dataset
(
dataset_root
=
args
.
dataset_root
,
transforms
=
eval_transforms
,
mode
=
'val'
)
if
args
.
model_name
not
in
MODELS
:
if
args
.
model_name
not
in
MODELS
:
raise
Exception
(
raise
Exception
(
'
--model_name
is invalid. it should be one of {}'
.
format
(
'
`--model_name`
is invalid. it should be one of {}'
.
format
(
str
(
list
(
MODELS
.
keys
()))))
str
(
list
(
MODELS
.
keys
()))))
model
=
MODELS
[
args
.
model_name
](
num_classes
=
train_dataset
.
num_classes
)
model
=
MODELS
[
args
.
model_name
](
num_classes
=
train_dataset
.
num_classes
)
...
@@ -176,7 +177,8 @@ def main(args):
...
@@ -176,7 +177,8 @@ def main(args):
lr_decay
,
lr_decay
,
momentum
=
0.9
,
momentum
=
0.9
,
parameter_list
=
model
.
parameters
(),
parameter_list
=
model
.
parameters
(),
regularization
=
fluid
.
regularizer
.
L2Decay
(
regularization_coeff
=
5e-4
))
regularization
=
fluid
.
regularizer
.
L2Decay
(
regularization_coeff
=
4e-5
))
train
(
train
(
model
,
model
,
train_dataset
,
train_dataset
,
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录