Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleRec
提交
0244a9db
P
PaddleRec
项目概览
PaddlePaddle
/
PaddleRec
通知
68
Star
12
Fork
5
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
27
列表
看板
标记
里程碑
合并请求
10
Wiki
1
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleRec
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
27
Issue
27
列表
看板
标记
里程碑
合并请求
10
合并请求
10
Pages
分析
分析
仓库分析
DevOps
Wiki
1
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
0244a9db
编写于
6月 17, 2020
作者:
C
chengmo
浏览文件
操作
浏览文件
下载
差异文件
Merge remote-tracking branch 'chengmo/fix_setup_print' into fix_setup_print
上级
191b52aa
707ca42d
变更
39
显示空白变更内容
内联
并排
Showing
39 changed file
with
138 addition
and
119 deletion
+138
-119
core/trainers/framework/dataset.py
core/trainers/framework/dataset.py
+3
-2
core/utils/dataloader_instance.py
core/utils/dataloader_instance.py
+7
-3
core/utils/envs.py
core/utils/envs.py
+22
-6
core/utils/util.py
core/utils/util.py
+33
-3
models/contentunderstanding/readme.md
models/contentunderstanding/readme.md
+2
-2
models/demo/movie_recommand/rank/config.yaml
models/demo/movie_recommand/rank/config.yaml
+1
-2
models/demo/movie_recommand/recall/config.yaml
models/demo/movie_recommand/recall/config.yaml
+1
-2
models/match/dssm/config.yaml
models/match/dssm/config.yaml
+1
-3
models/match/multiview-simnet/config.yaml
models/match/multiview-simnet/config.yaml
+1
-3
models/match/readme.md
models/match/readme.md
+1
-1
models/multitask/esmm/config.yaml
models/multitask/esmm/config.yaml
+35
-36
models/multitask/mmoe/config.yaml
models/multitask/mmoe/config.yaml
+1
-2
models/multitask/readme.md
models/multitask/readme.md
+1
-1
models/multitask/share-bottom/config.yaml
models/multitask/share-bottom/config.yaml
+1
-2
models/rank/afm/config.yaml
models/rank/afm/config.yaml
+1
-2
models/rank/dcn/config.yaml
models/rank/dcn/config.yaml
+1
-1
models/rank/deep_crossing/config.yaml
models/rank/deep_crossing/config.yaml
+1
-2
models/rank/deepfm/config.yaml
models/rank/deepfm/config.yaml
+1
-2
models/rank/din/config.yaml
models/rank/din/config.yaml
+1
-2
models/rank/ffm/config.yaml
models/rank/ffm/config.yaml
+1
-2
models/rank/fgcnn/config.yaml
models/rank/fgcnn/config.yaml
+1
-2
models/rank/fm/config.yaml
models/rank/fm/config.yaml
+1
-2
models/rank/fnn/config.yaml
models/rank/fnn/config.yaml
+1
-2
models/rank/logistic_regression/config.yaml
models/rank/logistic_regression/config.yaml
+1
-2
models/rank/nfm/config.yaml
models/rank/nfm/config.yaml
+1
-2
models/rank/pnn/config.yaml
models/rank/pnn/config.yaml
+1
-2
models/rank/readme.md
models/rank/readme.md
+1
-1
models/rank/wide_deep/config.yaml
models/rank/wide_deep/config.yaml
+1
-2
models/rank/xdeepfm/config.yaml
models/rank/xdeepfm/config.yaml
+1
-2
models/recall/gnn/config.yaml
models/recall/gnn/config.yaml
+1
-3
models/recall/gru4rec/config.yaml
models/recall/gru4rec/config.yaml
+1
-2
models/recall/ncf/config.yaml
models/recall/ncf/config.yaml
+1
-2
models/recall/readme.md
models/recall/readme.md
+1
-1
models/recall/ssr/config.yaml
models/recall/ssr/config.yaml
+1
-2
models/recall/word2vec/config.yaml
models/recall/word2vec/config.yaml
+1
-3
models/rerank/listwise/config.yaml
models/rerank/listwise/config.yaml
+1
-2
models/treebased/tdm/config.yaml
models/treebased/tdm/config.yaml
+1
-3
run.py
run.py
+4
-4
tools/build_script.sh
tools/build_script.sh
+1
-1
未找到文件。
core/trainers/framework/dataset.py
浏览文件 @
0244a9db
...
...
@@ -15,13 +15,13 @@
from
__future__
import
print_function
import
os
import
warnings
import
paddle.fluid
as
fluid
from
paddlerec.core.utils
import
envs
from
paddlerec.core.utils
import
dataloader_instance
from
paddlerec.core.reader
import
SlotReader
from
paddlerec.core.trainer
import
EngineMode
from
paddlerec.core.utils.util
import
split_files
__all__
=
[
"DatasetBase"
,
"DataLoader"
,
"QueueDataset"
]
...
...
@@ -123,7 +123,8 @@ class QueueDataset(DatasetBase):
for
x
in
os
.
listdir
(
train_data_path
)
]
if
context
[
"engine"
]
==
EngineMode
.
LOCAL_CLUSTER
:
file_list
=
context
[
"fleet"
].
split_files
(
file_list
)
file_list
=
split_files
(
file_list
,
context
[
"fleet"
].
worker_index
(),
context
[
"fleet"
].
worker_num
())
dataset
.
set_filelist
(
file_list
)
for
model_dict
in
context
[
"phases"
]:
...
...
core/utils/dataloader_instance.py
浏览文件 @
0244a9db
...
...
@@ -19,6 +19,7 @@ from paddlerec.core.utils.envs import get_global_env
from
paddlerec.core.utils.envs
import
get_runtime_environ
from
paddlerec.core.reader
import
SlotReader
from
paddlerec.core.trainer
import
EngineMode
from
paddlerec.core.utils.util
import
split_files
def
dataloader_by_name
(
readerclass
,
...
...
@@ -39,7 +40,8 @@ def dataloader_by_name(readerclass,
files
=
[
str
(
data_path
)
+
"/%s"
%
x
for
x
in
os
.
listdir
(
data_path
)]
if
context
[
"engine"
]
==
EngineMode
.
LOCAL_CLUSTER
:
files
=
context
[
"fleet"
].
split_files
(
files
)
files
=
split_files
(
files
,
context
[
"fleet"
].
worker_index
(),
context
[
"fleet"
].
worker_num
())
print
(
"file_list : {}"
.
format
(
files
))
reader
=
reader_class
(
yaml_file
)
...
...
@@ -80,7 +82,8 @@ def slotdataloader_by_name(readerclass, dataset_name, yaml_file, context):
files
=
[
str
(
data_path
)
+
"/%s"
%
x
for
x
in
os
.
listdir
(
data_path
)]
if
context
[
"engine"
]
==
EngineMode
.
LOCAL_CLUSTER
:
files
=
context
[
"fleet"
].
split_files
(
files
)
files
=
split_files
(
files
,
context
[
"fleet"
].
worker_index
(),
context
[
"fleet"
].
worker_num
())
print
(
"file_list: {}"
.
format
(
files
))
sparse
=
get_global_env
(
name
+
"sparse_slots"
,
"#"
)
...
...
@@ -133,7 +136,8 @@ def slotdataloader(readerclass, train, yaml_file, context):
files
=
[
str
(
data_path
)
+
"/%s"
%
x
for
x
in
os
.
listdir
(
data_path
)]
if
context
[
"engine"
]
==
EngineMode
.
LOCAL_CLUSTER
:
files
=
context
[
"fleet"
].
split_files
(
files
)
files
=
split_files
(
files
,
context
[
"fleet"
].
worker_index
(),
context
[
"fleet"
].
worker_num
())
print
(
"file_list: {}"
.
format
(
files
))
sparse
=
get_global_env
(
"sparse_slots"
,
"#"
,
namespace
)
...
...
core/utils/envs.py
浏览文件 @
0244a9db
...
...
@@ -18,7 +18,9 @@ import copy
import
os
import
socket
import
sys
import
six
import
traceback
import
six
global_envs
=
{}
global_envs_flatten
=
{}
...
...
@@ -101,6 +103,12 @@ def set_global_envs(envs):
name
=
"."
.
join
([
"dataset"
,
dataset
[
"name"
],
"type"
])
global_envs
[
name
]
=
"DataLoader"
if
get_platform
()
==
"LINUX"
and
six
.
PY3
:
print
(
"QueueDataset can not support PY3, change to DataLoader"
)
for
dataset
in
envs
[
"dataset"
]:
name
=
"."
.
join
([
"dataset"
,
dataset
[
"name"
],
"type"
])
global_envs
[
name
]
=
"DataLoader"
def
get_global_env
(
env_name
,
default_value
=
None
,
namespace
=
None
):
"""
...
...
@@ -253,11 +261,19 @@ def load_yaml(config):
use_full_loader
=
False
if
os
.
path
.
isfile
(
config
):
if
six
.
PY2
:
with
open
(
config
,
'r'
)
as
rb
:
if
use_full_loader
:
_config
=
yaml
.
load
(
rb
.
read
(),
Loader
=
yaml
.
FullLoader
)
else
:
_config
=
yaml
.
load
(
rb
.
read
())
return
_config
else
:
with
open
(
config
,
'r'
,
encoding
=
"utf-8"
)
as
rb
:
if
use_full_loader
:
_config
=
yaml
.
load
(
rb
.
read
(),
Loader
=
yaml
.
FullLoader
)
else
:
_config
=
yaml
.
load
(
rb
.
read
())
return
_config
else
:
raise
ValueError
(
"config {} can not be supported"
.
format
(
config
))
core/utils/util.py
浏览文件 @
0244a9db
...
...
@@ -19,11 +19,8 @@ import time
import
numpy
as
np
from
paddle
import
fluid
from
paddlerec.core.utils
import
fs
as
fs
def
save_program_proto
(
path
,
program
=
None
):
if
program
is
None
:
_program
=
fluid
.
default_main_program
()
else
:
...
...
@@ -171,6 +168,39 @@ def print_cost(cost, params):
return
log_str
def
split_files
(
files
,
trainer_id
,
trainers
):
"""
split files before distributed training,
example 1: files is [a, b, c ,d, e] and trainer_num = 2, then trainer
0 gets [a, b, c] and trainer 1 gets [d, e].
example 2: files is [a, b], and trainer_num = 3, then trainer 0 gets
[a], trainer 1 gets [b], trainer 2 gets []
Args:
files(list): file list need to be read.
Returns:
list: files belongs to this worker.
"""
if
not
isinstance
(
files
,
list
):
raise
TypeError
(
"files should be a list of file need to be read."
)
remainder
=
len
(
files
)
%
trainers
blocksize
=
int
(
len
(
files
)
/
trainers
)
blocks
=
[
blocksize
]
*
trainers
for
i
in
range
(
remainder
):
blocks
[
i
]
+=
1
trainer_files
=
[[]]
*
trainers
begin
=
0
for
i
in
range
(
trainers
):
trainer_files
[
i
]
=
files
[
begin
:
begin
+
blocks
[
i
]]
begin
+=
blocks
[
i
]
return
trainer_files
[
trainer_id
]
class
CostPrinter
(
object
):
"""
For count cost time && print cost log
...
...
models/contentunderstanding/readme.md
浏览文件 @
0244a9db
...
...
@@ -86,7 +86,7 @@ python -m paddlerec.run -m ./config.yaml # 自定义修改超参后,指定配
# 修改对应模型的config.yaml, workspace配置为当前目录的绝对路径
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: train_runner -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 修改phase阶段为infer的配置,参照config注释
# 修改完config.yaml后 执行:
...
...
@@ -106,7 +106,7 @@ python -m paddlerec.run -m ./config.yaml # 自定义修改超参后,指定配
# 修改对应模型的config.yaml, workspace配置为当前目录的绝对路径
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: train_runner -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 修改phase阶段为infer的配置,参照config注释
# 修改完config.yaml后 执行:
...
...
models/demo/movie_recommand/rank/config.yaml
浏览文件 @
0244a9db
...
...
@@ -64,8 +64,7 @@ runner:
device
:
cpu
-
name
:
runner_infer
epochs
:
1
class
:
single_infer
class
:
infer
print_interval
:
10000
init_model_path
:
"
increment/9"
# load model path
...
...
models/demo/movie_recommand/recall/config.yaml
浏览文件 @
0244a9db
...
...
@@ -64,8 +64,7 @@ runner:
device
:
cpu
-
name
:
runner_infer
epochs
:
1
class
:
single_infer
class
:
infer
print_interval
:
10000
init_model_path
:
"
increment/9"
# load model path
...
...
models/match/dssm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -56,9 +56,7 @@ runner:
init_model_path
:
"
"
# load model path
print_interval
:
2
-
name
:
infer_runner
class
:
single_infer
# num of epochs
epochs
:
1
class
:
infer
# device to run training or infer
device
:
cpu
print_interval
:
1
...
...
models/match/multiview-simnet/config.yaml
浏览文件 @
0244a9db
...
...
@@ -63,9 +63,7 @@ runner:
init_model_path
:
"
"
# load model path
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
# num of epochs
epochs
:
1
class
:
infer
# device to run training or infer
device
:
cpu
print_interval
:
1
...
...
models/match/readme.md
浏览文件 @
0244a9db
...
...
@@ -43,7 +43,7 @@ python -m paddlerec.run -m paddlerec.models.match.multiview-simnet # multiview-s
# 修改对应模型的config.yaml, workspace配置为当前目录的绝对路径
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: train_runner -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 修改phase阶段为infer的配置,参照config注释
# 修改完config.yaml后 执行:
...
...
models/multitask/esmm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -16,12 +16,12 @@
workspace
:
"
paddlerec.models.multitask.esmm"
dataset
:
-
name
:
dataset_train
-
name
:
dataset_train
batch_size
:
1
type
:
QueueDataset
data_path
:
"
{workspace}/data/train"
data_converter
:
"
{workspace}/esmm_reader.py"
-
name
:
dataset_infer
-
name
:
dataset_infer
batch_size
:
1
type
:
QueueDataset
data_path
:
"
{workspace}/data/test"
...
...
@@ -39,7 +39,7 @@ hyper_parameters:
mode
:
[
train_runner
,
infer_runner
]
runner
:
-
name
:
train_runner
-
name
:
train_runner
class
:
train
device
:
cpu
epochs
:
3
...
...
@@ -49,20 +49,19 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
10
phases
:
[
train
]
-
name
:
infer_runner
-
name
:
infer_runner
class
:
infer
init_model_path
:
"
increment/0
"
init_model_path
:
"
increment/1
"
device
:
cpu
epochs
:
1
print_interval
:
1
phases
:
[
infer
]
phase
:
-
name
:
train
-
name
:
train
model
:
"
{workspace}/model.py"
dataset_name
:
dataset_train
thread_num
:
1
-
name
:
infer
-
name
:
infer
model
:
"
{workspace}/model.py"
dataset_name
:
dataset_infer
thread_num
:
1
models/multitask/mmoe/config.yaml
浏览文件 @
0244a9db
...
...
@@ -52,10 +52,9 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
10
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
3
phase
:
-
name
:
train
...
...
models/multitask/readme.md
浏览文件 @
0244a9db
...
...
@@ -77,7 +77,7 @@ python -m paddlerec.run -m ./config.yaml # 自定义修改超参后,指定配
# 修改对应模型的config.yaml, workspace配置为当前目录的绝对路径
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: train_runner -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 修改phase阶段为infer的配置,参照config注释
# 修改完config.yaml后 执行:
...
...
models/multitask/share-bottom/config.yaml
浏览文件 @
0244a9db
...
...
@@ -51,10 +51,9 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
5
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
3
phase
:
-
name
:
train
...
...
models/rank/afm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -59,8 +59,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/dcn/config.yaml
浏览文件 @
0244a9db
...
...
@@ -60,7 +60,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
epochs
:
1
device
:
cpu
init_model_path
:
"
increment/0"
...
...
models/rank/deep_crossing/config.yaml
浏览文件 @
0244a9db
...
...
@@ -58,8 +58,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/deepfm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -58,8 +58,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/din/config.yaml
浏览文件 @
0244a9db
...
...
@@ -56,8 +56,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/ffm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -57,8 +57,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/fgcnn/config.yaml
浏览文件 @
0244a9db
...
...
@@ -62,8 +62,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/fm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -57,8 +57,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/fnn/config.yaml
浏览文件 @
0244a9db
...
...
@@ -68,8 +68,7 @@ runner:
save_inference_path
:
"
inference_fnn"
print_interval
:
1
-
name
:
infer_runner
trainer_class
:
single_infer
epochs
:
1
trainer_class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/logistic_regression/config.yaml
浏览文件 @
0244a9db
...
...
@@ -56,8 +56,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/nfm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -63,8 +63,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/pnn/config.yaml
浏览文件 @
0244a9db
...
...
@@ -60,8 +60,7 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
print_interval
:
1
...
...
models/rank/readme.md
浏览文件 @
0244a9db
...
...
@@ -98,7 +98,7 @@ python -m paddlerec.run -m ./config.yaml
```
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: runner1 -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 如果训练阶段和预测阶段的模型输入一致,phase不需要改动,复用train的即可
# 修改完config.yaml后 执行:
...
...
models/rank/wide_deep/config.yaml
浏览文件 @
0244a9db
...
...
@@ -54,8 +54,7 @@ runner:
save_checkpoint_path
:
"
increment"
save_inference_path
:
"
inference"
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
...
...
models/rank/xdeepfm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -55,8 +55,7 @@ runner:
save_checkpoint_path
:
"
increment"
save_inference_path
:
"
inference"
-
name
:
infer_runner
class
:
single_infer
epochs
:
1
class
:
infer
device
:
cpu
init_model_path
:
"
increment/0"
...
...
models/recall/gnn/config.yaml
浏览文件 @
0244a9db
...
...
@@ -61,9 +61,7 @@ runner:
init_model_path
:
"
"
# load model path
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
# num of epochs
epochs
:
1
class
:
infer
# device to run training or infer
device
:
cpu
print_interval
:
1
...
...
models/recall/gru4rec/config.yaml
浏览文件 @
0244a9db
...
...
@@ -54,10 +54,9 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
10
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
1
phase
:
-
name
:
train
...
...
models/recall/ncf/config.yaml
浏览文件 @
0244a9db
...
...
@@ -51,10 +51,9 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
10
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
1
phase
:
-
name
:
train
...
...
models/recall/readme.md
浏览文件 @
0244a9db
...
...
@@ -95,7 +95,7 @@ python -m paddlerec.run -m ./config.yaml # 自定义修改超参后,指定配
# 修改对应模型的config.yaml, workspace配置为当前目录的绝对路径
# 修改对应模型的config.yaml,mode配置infer_runner
# 示例: mode: train_runner -> mode: infer_runner
# infer_runner中 class配置为 class:
single_
infer
# infer_runner中 class配置为 class: infer
# 修改phase阶段为infer的配置,参照config注释
# 修改完config.yaml后 执行:
...
...
models/recall/ssr/config.yaml
浏览文件 @
0244a9db
...
...
@@ -50,10 +50,9 @@ runner:
save_inference_path
:
"
inference"
print_interval
:
10
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
1
phase
:
-
name
:
train
...
...
models/recall/word2vec/config.yaml
浏览文件 @
0244a9db
...
...
@@ -61,9 +61,7 @@ runner:
init_model_path
:
"
"
# load model path
print_interval
:
1
-
name
:
infer_runner
class
:
single_infer
# num of epochs
epochs
:
1
class
:
infer
# device to run training or infer
device
:
cpu
init_model_path
:
"
increment/0"
# load model path
...
...
models/rerank/listwise/config.yaml
浏览文件 @
0244a9db
...
...
@@ -51,10 +51,9 @@ runner:
save_checkpoint_path
:
"
increment"
save_inference_path
:
"
inference"
-
name
:
infer_runner
class
:
single_
infer
class
:
infer
init_model_path
:
"
increment/0"
device
:
cpu
epochs
:
3
phase
:
-
name
:
train
...
...
models/treebased/tdm/config.yaml
浏览文件 @
0244a9db
...
...
@@ -80,10 +80,8 @@ runner:
print_interval
:
10
-
name
:
runner2
class
:
single_
infer
class
:
infer
startup_class_path
:
"
{workspace}/tdm_startup.py"
# num of epochs
epochs
:
1
# device to run training or infer
device
:
cpu
init_model_path
:
"
increment/0"
# load model path
...
...
run.py
浏览文件 @
0244a9db
...
...
@@ -139,7 +139,7 @@ def get_engine(args, running_config, mode):
engine
=
"LOCAL_CLUSTER_TRAIN"
if
engine
not
in
engine_choices
:
raise
ValueError
(
"{} can
not
be chosen in {}"
.
format
(
engine_class
,
raise
ValueError
(
"{} can
only
be chosen in {}"
.
format
(
engine_class
,
engine_choices
))
run_engine
=
engines
[
transpiler
].
get
(
engine
,
None
)
...
...
@@ -439,8 +439,8 @@ def local_cluster_engine(args):
if
fleet_mode
==
"COLLECTIVE"
:
cluster_envs
[
"selected_gpus"
]
=
selected_gpus
gpus
=
selected_gpus
.
split
(
","
)
gpu
_num
=
get_worker_num
(
run_extras
,
len
(
gpus
))
cluster_envs
[
"selected_gpus"
]
=
','
.
join
(
gpus
[:
gpu
_num
])
worker
_num
=
get_worker_num
(
run_extras
,
len
(
gpus
))
cluster_envs
[
"selected_gpus"
]
=
','
.
join
(
gpus
[:
worker
_num
])
cluster_envs
[
"server_num"
]
=
server_num
cluster_envs
[
"worker_num"
]
=
worker_num
...
...
tools/build_script.sh
浏览文件 @
0244a9db
...
...
@@ -49,7 +49,7 @@ function model_test() {
root_dir
=
`
pwd
`
all_model
=
$(
find
${
root_dir
}
-name
config.yaml
)
special_models
=(
"demo"
"pnn"
"fgcnn"
"
esmm
"
)
special_models
=(
"demo"
"pnn"
"fgcnn"
"
gru4rec"
"tagspace
"
)
for
model
in
${
all_model
}
do
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录