Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
c4a6a0e2
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2298
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
c4a6a0e2
编写于
3月 18, 2020
作者:
T
tangwei12
提交者:
GitHub
3月 18, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Revert "Integrated API of Parameter Server (#22710)" test=develop (#23071)
This reverts commit
66fce9e8
.
上级
880eb04d
变更
17
隐藏空白更改
内联
并排
Showing
17 changed file
with
56 addition
and
183 deletion
+56
-183
python/paddle/fluid/incubate/__init__.py
python/paddle/fluid/incubate/__init__.py
+1
-2
python/paddle/fluid/incubate/fleet/__init__.py
python/paddle/fluid/incubate/fleet/__init__.py
+1
-4
python/paddle/fluid/incubate/fleet/base/__init__.py
python/paddle/fluid/incubate/fleet/base/__init__.py
+0
-6
python/paddle/fluid/incubate/fleet/base/fleet_base.py
python/paddle/fluid/incubate/fleet/base/fleet_base.py
+2
-2
python/paddle/fluid/incubate/fleet/base/role_maker.py
python/paddle/fluid/incubate/fleet/base/role_maker.py
+5
-5
python/paddle/fluid/incubate/fleet/parameter_server/__init__.py
.../paddle/fluid/incubate/fleet/parameter_server/__init__.py
+0
-17
python/paddle/fluid/incubate/fleet/parameter_server/distribute_transpiler/__init__.py
.../fleet/parameter_server/distribute_transpiler/__init__.py
+1
-1
python/paddle/fluid/incubate/fleet/parameter_server/distribute_transpiler/distributed_strategy.py
...eter_server/distribute_transpiler/distributed_strategy.py
+0
-15
python/paddle/fluid/incubate/fleet/parameter_server/pslib/__init__.py
...e/fluid/incubate/fleet/parameter_server/pslib/__init__.py
+16
-28
python/paddle/fluid/incubate/fleet/parameter_server/pslib/node.py
...addle/fluid/incubate/fleet/parameter_server/pslib/node.py
+0
-1
python/paddle/fluid/tests/unittests/CMakeLists.txt
python/paddle/fluid/tests/unittests/CMakeLists.txt
+0
-2
python/paddle/fluid/tests/unittests/test_communicator_sync.py
...on/paddle/fluid/tests/unittests/test_communicator_sync.py
+1
-1
python/paddle/fluid/tests/unittests/test_dataset.py
python/paddle/fluid/tests/unittests/test_dataset.py
+27
-2
python/paddle/fluid/tests/unittests/test_dist_fleet_base.py
python/paddle/fluid/tests/unittests/test_dist_fleet_base.py
+1
-1
python/paddle/fluid/tests/unittests/test_distributed_strategy.py
...paddle/fluid/tests/unittests/test_distributed_strategy.py
+1
-1
python/paddle/fluid/tests/unittests/test_fleet_input_with_psilib.py
...dle/fluid/tests/unittests/test_fleet_input_with_psilib.py
+0
-65
python/setup.py.in
python/setup.py.in
+0
-30
未找到文件。
python/paddle/fluid/incubate/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -14,5 +14,4 @@
# incubate directory is mainly for internal use
# after we have tested incubate APIs in industrial application for a period
# we will move stable functions into fluid
from
__future__
import
print_function
__version__
=
'0.1.0'
python/paddle/fluid/incubate/fleet/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -11,7 +11,4 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
from
__future__
import
print_function
from
.
import
base
from
.base
import
*
__version__
=
'0.1.0'
python/paddle/fluid/incubate/fleet/base/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -10,9 +10,3 @@
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
from
.
import
fleet_base
from
.fleet_base
import
*
from
.
import
role_maker
from
.role_maker
import
*
python/paddle/fluid/incubate/fleet/base/fleet_base.py
浏览文件 @
c4a6a0e2
...
...
@@ -16,7 +16,7 @@ from __future__ import print_function
import
abc
from
paddle.fluid.core
import
CPUPlace
import
paddle.fluid
as
fluid
from
paddle.fluid.executor
import
Executor
from
paddle.fluid.optimizer
import
SGD
...
...
@@ -193,7 +193,7 @@ class Fleet(object):
Returns:
None
"""
self
.
_executor
=
Executor
(
CPUPlace
())
self
.
_executor
=
Executor
(
fluid
.
CPUPlace
())
if
role_maker
and
not
isinstance
(
role_maker
,
RoleMakerBase
):
raise
TypeError
(
"role_maker must be an instance of RoleMakerBase"
)
...
...
python/paddle/fluid/incubate/fleet/base/role_maker.py
浏览文件 @
c4a6a0e2
...
...
@@ -14,9 +14,9 @@
"""Defination of Role Makers."""
from
__future__
import
print_function
import
paddle.fluid
as
fluid
import
os
from
paddle.fluid.core
import
Gloo
import
time
__all__
=
[
'Role'
,
'RoleMakerBase'
,
'MPISymetricRoleMaker'
,
'UserDefinedRoleMaker'
,
...
...
@@ -577,7 +577,7 @@ class GeneralRoleMaker(RoleMakerBase):
current_id
=
int
(
os
.
environ
[
"PADDLE_TRAINER_ID"
])
self
.
_node_type
=
1
self
.
_cur_endpoint
=
worker_endpoints
[
current_id
]
gloo
=
Gloo
()
gloo
=
fluid
.
core
.
Gloo
()
gloo
.
init
(
current_id
,
len
(
worker_endpoints
),
self
.
_hdfs_path
.
rstrip
(
"/"
)
+
"/trainer"
,
...
...
@@ -597,7 +597,7 @@ class GeneralRoleMaker(RoleMakerBase):
current_id
=
eplist
.
index
(
cur_endpoint
)
self
.
_node_type
=
0
self
.
_cur_endpoint
=
cur_endpoint
gloo
=
Gloo
()
gloo
=
fluid
.
core
.
Gloo
()
gloo
.
init
(
current_id
,
len
(
eplist
),
self
.
_hdfs_path
.
rstrip
(
"/"
)
+
"/pserver"
,
...
...
@@ -605,7 +605,7 @@ class GeneralRoleMaker(RoleMakerBase):
self
.
_prefix
)
self
.
_node_type_comm
=
gloo
gloo
=
Gloo
()
gloo
=
fluid
.
core
.
Gloo
()
all_list
=
worker_endpoints
+
eplist
gloo
.
init
(
all_list
.
index
(
self
.
_cur_endpoint
),
...
...
python/paddle/fluid/incubate/fleet/parameter_server/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -11,20 +11,3 @@
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
from
paddle.fluid.incubate.fleet.parameter_server
import
version
from
paddle.fluid.incubate.fleet.parameter_server.distributed_strategy
import
TrainerRuntimeConfig
from
paddle.fluid.incubate.fleet.parameter_server.distributed_strategy
import
StrategyFactory
__all__
=
[
'TrainerRuntimeConfig'
,
'StrategyFactory'
,
'fleet'
]
fleet
=
None
if
version
.
is_transpiler
():
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
fleet
as
fleet_transpiler
fleet
=
fleet_transpiler
else
:
from
paddle.fluid.incubate.fleet.parameter_server.pslib
import
fleet
as
fleet_pslib
fleet
=
fleet_pslib
python/paddle/fluid/incubate/fleet/parameter_server/distribute_transpiler/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -27,7 +27,7 @@ from paddle.fluid.executor import Executor
from
paddle.fluid.parallel_executor
import
ParallelExecutor
from
paddle.fluid.optimizer
import
Optimizer
from
paddle.fluid.incubate.fleet.parameter_server.distributed_strategy
import
TrainerRuntimeConfig
,
DistributedStrategy
,
SyncStrategy
,
AsyncStrategy
,
HalfAsyncStrategy
,
GeoStrategy
,
StrategyFactory
from
paddle.fluid.incubate.fleet.parameter_server.distribute
_transpiler.distribute
d_strategy
import
TrainerRuntimeConfig
,
DistributedStrategy
,
SyncStrategy
,
AsyncStrategy
,
HalfAsyncStrategy
,
GeoStrategy
,
StrategyFactory
from
paddle.fluid.transpiler.distribute_transpiler
import
DistributeTranspiler
as
OriginTranspiler
from
paddle.fluid.transpiler.distribute_transpiler
import
DistributeTranspilerConfig
,
ServerRuntimeConfig
,
DistributedMode
...
...
python/paddle/fluid/incubate/fleet/parameter_server/distributed_strategy.py
→
python/paddle/fluid/incubate/fleet/parameter_server/distribute
_transpiler/distribute
d_strategy.py
浏览文件 @
c4a6a0e2
...
...
@@ -124,19 +124,10 @@ class TrainerRuntimeConfig(object):
return
self
.
display
(
self
.
get_communicator_flags
())
class
PSLibRuntimeConfig
(
object
):
def
__init__
(
self
):
self
.
runtime_configs
=
{}
def
get_runtime_configs
(
self
):
return
self
.
runtime_configs
class
DistributedStrategy
(
object
):
def
__init__
(
self
):
self
.
_program_config
=
DistributeTranspilerConfig
()
self
.
_trainer_runtime_config
=
TrainerRuntimeConfig
()
self
.
_pslib_runtime_config
=
PSLibRuntimeConfig
()
self
.
_server_runtime_config
=
ServerRuntimeConfig
()
num_threads
=
int
(
os
.
getenv
(
"CPU_NUM"
,
"1"
))
...
...
@@ -213,12 +204,6 @@ class DistributedStrategy(object):
"check_trainer_runtime_config must be implemented by derived class. You should use StrategyFactory to create DistributedStrategy."
)
def
get_pslib_runtime_config
(
self
):
return
self
.
_pslib_runtime_config
def
set_pslib_runtime_config
(
self
,
config
):
self
.
_pslib_runtime_config
.
runtime_configs
=
config
def
get_server_runtime_config
(
self
):
return
self
.
_server_runtime_config
...
...
python/paddle/fluid/incubate/fleet/parameter_server/pslib/__init__.py
浏览文件 @
c4a6a0e2
...
...
@@ -18,13 +18,11 @@ from .optimizer_factory import *
from
google.protobuf
import
text_format
import
paddle.fluid
as
fluid
from
paddle.fluid.framework
import
Program
from
paddle.fluid.optimizer
import
Optimizer
from
paddle.fluid.incubate.fleet.base.fleet_base
import
Fleet
from
paddle.fluid.incubate.fleet.base.fleet_base
import
Mode
from
paddle.fluid.incubate.fleet.base.fleet_base
import
DistributedOptimizer
from
paddle.fluid.incubate.fleet.base.role_maker
import
MPISymetricRoleMaker
from
paddle.fluid.incubate.fleet.parameter_server.distributed_strategy
import
TrainerRuntimeConfig
,
DistributedStrategy
,
SyncStrategy
,
AsyncStrategy
,
HalfAsyncStrategy
,
GeoStrategy
,
StrategyFactory
class
PSLib
(
Fleet
):
...
...
@@ -224,32 +222,7 @@ class PSLib(Fleet):
optimizer(DownpourOptimizer): downpour optimizer
"""
if
not
isinstance
(
optimizer
,
Optimizer
):
raise
ValueError
(
"optimizer must be an instance of Optimizer"
)
if
not
fleet
.
_is_initialized
:
raise
ValueError
(
"use fleet.init(role) to initialize the role of current node before optimizer.minimize(loss)"
)
if
strategy
:
if
isinstance
(
strategy
,
dict
):
self
.
_strategy
=
strategy
elif
isinstance
(
strategy
,
DistributedStrategy
):
if
isinstance
(
strategy
,
AsyncStrategy
):
self
.
_strategy
=
strategy
.
get_pslib_runtime_config
(
).
get_runtime_configs
()
else
:
raise
TypeError
(
"In {} mode, strategy must be an instance of AsyncStrategy, or Dict"
.
format
(
fleet
.
_mode
))
else
:
raise
TypeError
(
"In {} mode, strategy must be an instance of AsyncStrategy, or Dict"
.
format
(
fleet
.
_mode
))
else
:
self
.
_strategy
=
{}
self
.
_optimizer
=
DownpourOptimizer
(
optimizer
,
self
.
_strategy
)
self
.
_optimizer
=
DownpourOptimizer
(
optimizer
,
strategy
)
return
self
.
_optimizer
def
save_inference_model
(
self
,
...
...
@@ -444,6 +417,21 @@ class PSLib(Fleet):
self
.
_fleet_ptr
.
clear_model
()
self
.
_role_maker
.
_barrier_worker
()
def
clear_model
(
self
):
"""
clear_model() will be called by user. It will clear sparse model.
Examples:
.. code-block:: python
fleet.clear_model()
"""
self
.
_role_maker
.
_barrier_worker
()
if
self
.
_role_maker
.
is_first_worker
():
self
.
_fleet_ptr
.
clear_model
()
self
.
_role_maker
.
_barrier_worker
()
def
load_one_table
(
self
,
table_id
,
model_path
,
**
kwargs
):
"""
load pslib model for one table or load params from paddle model
...
...
python/paddle/fluid/incubate/fleet/parameter_server/pslib/node.py
浏览文件 @
c4a6a0e2
...
...
@@ -13,7 +13,6 @@
"""Defination of Server and Worker."""
from
.
import
ps_pb2
as
pslib
from
six.moves
import
reduce
class
Server
(
object
):
...
...
python/paddle/fluid/tests/unittests/CMakeLists.txt
浏览文件 @
c4a6a0e2
...
...
@@ -28,7 +28,6 @@ list(APPEND MIXED_DIST_TEST_OPS test_communicator_geo)
list
(
APPEND MIXED_DIST_TEST_OPS test_communicator_half_async
)
list
(
APPEND MIXED_DIST_TEST_OPS test_communicator_sync
)
list
(
APPEND MIXED_DIST_TEST_OPS test_fleet_api_input
)
list
(
APPEND MIXED_DIST_TEST_OPS test_fleet_input_with_psilib
)
foreach
(
TEST_OP
${
MIXED_DIST_TEST_OPS
}
)
list
(
REMOVE_ITEM TEST_OPS
${
TEST_OP
}
)
endforeach
()
...
...
@@ -283,7 +282,6 @@ if(WITH_DISTRIBUTE)
py_test_modules
(
test_nce_remote_table_op MODULES test_nce_remote_table_op ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_recv_save_op MODULES test_recv_save_op ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_transpiler_ops MODULES test_transpiler_ops ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_fleet_input_with_psilib MODULES test_fleet_input_with_psilib ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_communicator_async MODULES test_communicator_async ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_communicator_geo MODULES test_communicator_geo ENVS
${
dist_ENVS
}
)
py_test_modules
(
test_communicator_half_async MODULES test_communicator_half_async ENVS
${
dist_ENVS
}
FLAGS_communicator_send_queue_size=1 FLAGS_communicator_max_merge_var_num=1
)
...
...
python/paddle/fluid/tests/unittests/test_communicator_sync.py
浏览文件 @
c4a6a0e2
...
...
@@ -25,7 +25,7 @@ from paddle.fluid.communicator import Communicator
import
paddle.fluid.incubate.fleet.base.role_maker
as
role_maker
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
fleet
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
StrategyFactory
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
.distributed_strategy
import
StrategyFactory
class
TestCommunicator
(
unittest
.
TestCase
):
...
...
python/paddle/fluid/tests/unittests/test_dataset.py
浏览文件 @
c4a6a0e2
...
...
@@ -720,6 +720,33 @@ class TestDatasetWithFetchHandler(unittest.TestCase):
except
Exception
as
e
:
self
.
assertTrue
(
False
)
def
test_fetch_handler
(
self
):
"""
Test Dataset With Fetch Handler. TestCases.
"""
slots_vars
,
out
=
self
.
net
()
files
=
[
"test_queue_dataset_run_a.txt"
,
"test_queue_dataset_run_b.txt"
]
dataset
=
self
.
get_dataset
(
slots_vars
,
files
)
exe
=
fluid
.
Executor
(
fluid
.
CPUPlace
())
exe
.
run
(
fluid
.
default_startup_program
())
fh
=
fluid
.
executor
.
FetchHandler
(
out
.
name
)
fh
.
help
()
try
:
exe
.
train_from_dataset
(
program
=
fluid
.
default_main_program
(),
dataset
=
dataset
,
fetch_handler
=
fh
)
except
ImportError
as
e
:
print
(
"warning: we skip trainer_desc_pb2 import problem in windows"
)
except
RuntimeError
as
e
:
error_msg
=
"dataset is need and should be initialized"
self
.
assertEqual
(
error_msg
,
cpt
.
get_exception_message
(
e
))
except
Exception
as
e
:
self
.
assertTrue
(
False
)
class
TestDataset2
(
unittest
.
TestCase
):
""" TestCases for Dataset. """
...
...
@@ -769,7 +796,6 @@ class TestDataset2(unittest.TestCase):
print
(
"warning: no mpi4py"
)
adam
=
fluid
.
optimizer
.
Adam
(
learning_rate
=
0.000005
)
try
:
fleet
.
init
()
adam
=
fleet
.
distributed_optimizer
(
adam
)
adam
.
minimize
([
fake_cost
],
[
scope
])
except
AttributeError
as
e
:
...
...
@@ -832,7 +858,6 @@ class TestDataset2(unittest.TestCase):
print
(
"warning: no mpi4py"
)
adam
=
fluid
.
optimizer
.
Adam
(
learning_rate
=
0.000005
)
try
:
fleet
.
init
()
adam
=
fleet
.
distributed_optimizer
(
adam
,
strategy
=
{
...
...
python/paddle/fluid/tests/unittests/test_dist_fleet_base.py
浏览文件 @
c4a6a0e2
...
...
@@ -37,7 +37,7 @@ import paddle.fluid as fluid
import
paddle.fluid.incubate.fleet.base.role_maker
as
role_maker
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
fleet
from
paddle.fluid.transpiler.distribute_transpiler
import
DistributeTranspilerConfig
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
StrategyFactory
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
.distributed_strategy
import
StrategyFactory
__all__
=
[
'FleetDistRunnerBase'
,
'TestFleetBase'
,
'runtime_main'
]
...
...
python/paddle/fluid/tests/unittests/test_distributed_strategy.py
浏览文件 @
c4a6a0e2
...
...
@@ -15,7 +15,7 @@
import
unittest
import
paddle.fluid
as
fluid
from
paddle.fluid.transpiler.distribute_transpiler
import
DistributeTranspilerConfig
,
ServerRuntimeConfig
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
TrainerRuntimeConfig
,
StrategyFactory
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
.distributed_strategy
import
TrainerRuntimeConfig
,
StrategyFactory
from
paddle.fluid.incubate.fleet.parameter_server.distribute_transpiler
import
fleet
import
paddle.fluid.incubate.fleet.base.role_maker
as
role_maker
import
os
...
...
python/paddle/fluid/tests/unittests/test_fleet_input_with_psilib.py
已删除
100644 → 0
浏览文件 @
880eb04d
# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
import
unittest
from
six.moves
import
reduce
import
paddle.fluid
as
fluid
from
paddle.fluid.transpiler.distribute_transpiler
import
DistributeTranspilerConfig
from
paddle.fluid.incubate.fleet.base.role_maker
import
UserDefinedRoleMaker
from
paddle.fluid.incubate.fleet.base.role_maker
import
UserDefinedCollectiveRoleMaker
from
paddle.fluid.incubate.fleet.base.role_maker
import
Role
import
paddle.fluid.incubate.fleet.base.role_maker
as
role_maker
from
paddle.fluid.incubate.fleet.parameter_server.pslib
import
fleet
from
paddle.fluid.incubate.fleet.parameter_server.distributed_strategy
import
StrategyFactory
from
dist_simnet_bow
import
train_network
class
FleetPSLibTest
(
unittest
.
TestCase
):
def
test_transpile
(
self
):
role
=
role_maker
.
UserDefinedRoleMaker
(
current_id
=
0
,
role
=
role_maker
.
Role
.
SERVER
,
worker_num
=
2
,
server_endpoints
=
[
"127.0.0.1:36011"
,
"127.0.0.1:36012"
])
optimizer
=
fluid
.
optimizer
.
SGD
(
0.1
)
# case5
self
.
assertRaises
(
Exception
,
fleet
.
distributed_optimizer
,
optimizer
,
"Adam"
)
fleet
.
init
(
role
)
avg_cost
,
_
,
_
=
train_network
(
128
,
False
,
True
)
# case1
strategy
=
StrategyFactory
.
create_async_strategy
()
fleet
.
distributed_optimizer
(
optimizer
,
strategy
)
# case2
strategy
=
{}
fleet
.
distributed_optimizer
(
optimizer
,
strategy
)
# case3
self
.
assertRaises
(
Exception
,
fleet
.
distributed_optimizer
,
optimizer
,
"Adam"
)
# case4
self
.
assertRaises
(
Exception
,
fleet
.
distributed_optimizer
,
"Adam"
,
"Adam"
)
if
__name__
==
'__main__'
:
unittest
.
main
()
python/setup.py.in
浏览文件 @
c4a6a0e2
import subprocess
import os
import os.path
import errno
import re
import shutil
import sys
...
...
@@ -105,35 +104,6 @@ def mkl():
write_version_py(filename='@PADDLE_BINARY_DIR@/python/paddle/version.py')
def write_distributed_training_mode_py(filename='paddle/fluid/incubate/fleet/parameter_server/version.py'):
cnt = '''from __future__ import print_function
# THIS FILE IS GENERATED FROM PADDLEPADDLE SETUP.PY
from paddle.fluid.incubate.fleet.base.fleet_base import Mode
BUILD_MODE=Mode.%(mode)s
def is_transpiler():
return Mode.TRANSPILER == BUILD_MODE
'''
dirname = os.path.dirname(filename)
try:
os.makedirs(dirname)
except OSError as e:
if e.errno != errno.EEXIST:
raise
with open(filename, 'w') as f:
f.write(cnt % {
'mode': 'PSLIB' if '${WITH_PSLIB}' == 'ON' else 'TRANSPILER'
})
write_distributed_training_mode_py(filename='@PADDLE_BINARY_DIR@/python/paddle/fluid/incubate/fleet/parameter_server/version.py')
packages=['paddle',
'paddle.libs',
'paddle.utils',
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录