Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
d52586a9
P
Paddle
项目概览
PaddlePaddle
/
Paddle
1 年多 前同步成功
通知
2305
Star
20932
Fork
5423
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
d52586a9
编写于
3月 25, 2019
作者:
X
xjqbest
提交者:
dongdaxiang
3月 29, 2019
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
add doc string
test=develop
上级
6be9f719
变更
4
隐藏空白更改
内联
并排
Showing
4 changed file
with
110 addition
and
21 deletion
+110
-21
python/paddle/fluid/async_executor.py
python/paddle/fluid/async_executor.py
+27
-10
python/paddle/fluid/device_worker.py
python/paddle/fluid/device_worker.py
+52
-1
python/paddle/fluid/incubate/fleet/parameter_server/__init__.py
.../paddle/fluid/incubate/fleet/parameter_server/__init__.py
+4
-0
python/paddle/fluid/tests/unittests/test_dataset.py
python/paddle/fluid/tests/unittests/test_dataset.py
+27
-10
未找到文件。
python/paddle/fluid/async_executor.py
浏览文件 @
d52586a9
...
...
@@ -80,6 +80,11 @@ class AsyncExecutor(object):
def
__init__
(
self
,
place
=
None
,
run_mode
=
""
):
"""
Init.
Example:
>>> place = fluid.CPUPlace()
>>> async_executor = fluid.AsyncExecutor(place)
Args:
place(Place): CPUPlace or GPUPlace.
run_mode(str): default is empty string.
...
...
@@ -99,6 +104,14 @@ class AsyncExecutor(object):
def
run
(
self
,
program
,
data_feed
,
filelist
,
thread_num
,
fetch
,
debug
=
False
):
"""
Run program by this AsyncExecutor.
Example:
>>> place = fluid.CPUPlace()
>>> async_executor = fluid.AsyncExecutor(place)
>>> async_executor.run(default_main_program(),
my_data_feed_desc,
["a.txt", "b.txt"])
Args:
program(Program): the program that need to run, if not provied,
then default_main_program will be used.
...
...
@@ -235,12 +248,13 @@ class AsyncExecutor(object):
>>> exe.download_data("/xxx/xxx/xx/",
>>> "./data", "afs://
>>> xxx.xxx.xxx.xxx:9901", "xxx,yyy")
Args:
afs_path(str): afs_path defined by users
local_path(str): download data path
fs_default_name(str): file system server address
ugi(str): hadoop ugi
file_cn(int): a user can specify file number for debugging
file_cn
t
(int): a user can specify file number for debugging
hadoop_home(str): hadoop home path
process_num(int): download process num
"""
...
...
@@ -298,10 +312,11 @@ class AsyncExecutor(object):
def
init_server
(
self
,
dist_desc
):
"""
initialize server of current node if current process is a server
Initialize server of current node if current process is a server.
Args:
dist_desc(str): a protobuf string that describes
how to init a worker and a server
dist_desc(str): a protobuf string that describes
how to init a worker and a server
"""
if
self
.
instance
is
None
:
raise
ValueError
(
...
...
@@ -319,11 +334,12 @@ class AsyncExecutor(object):
def
init_worker
(
self
,
dist_desc
,
startup_program
):
"""
initialize worker of current node if current process is a worker
Initialize worker of current node if current process is a worker.
Args:
dist_desc(str): a protobuf string that describes
how to init a worker and a server
startup_program(fluid.Program): startup program of current process
dist_desc(str): a protobuf string that describes
how to init a worker and a server
startup_program(fluid.Program): startup program of current process
"""
if
self
.
instance
is
None
:
raise
ValueError
(
...
...
@@ -364,9 +380,10 @@ class AsyncExecutor(object):
def
save_model
(
self
,
save_path
):
"""
save_model command that can be invoked from one of the worker
model parameters are saved in servers and upload to save_path of file system
model parameters are saved in servers and upload to save_path of file system.
Args:
save_path(str): save path to file system
save_path(str): save path to file system
"""
if
self
.
instance
is
None
:
raise
ValueError
(
...
...
python/paddle/fluid/device_worker.py
浏览文件 @
d52586a9
...
...
@@ -17,32 +17,83 @@ __all__ = ['DeviceWorker', 'Hogwild', 'DownpourSGD']
class
DeviceWorker
(
object
):
"""
DeviceWorker is a abstract class, which generates worker desc.
"""
def
__init__
(
self
):
"""
Init.
"""
self
.
program_
=
None
def
set_fleet_desc
(
self
,
fleet_desc
):
"""
Set fleet desc.
Args:
fleet_desc(PSParameter): pslib.PSParameter object
"""
self
.
fleet_desc_
=
fleet_desc
def
set_program
(
self
,
program
):
"""
Set program.
Args:
program(Program): a Program object
"""
self
.
program_
=
program
def
gen_worker_desc
(
self
,
trainer_desc
):
pass
"""
Generator worker desc.
Args:
trainer_desc(TrainerDesc): a TrainerDesc object
"""
raise
NotImplementedError
(
"DeviceWorker does not implement gen_worker_desc, "
"please use Hogwild or DownpourSGD, etc."
)
class
Hogwild
(
DeviceWorker
):
"""
Hogwild is a kind of SGD algorithm.
"""
def
__init__
(
self
):
"""
Init.
"""
super
(
Hogwild
,
self
).
__init__
()
def
gen_worker_desc
(
self
,
trainer_desc
):
"""
Generator worker desc, which device worker is HogwildWorker.
Args:
trainer_desc(TrainerDesc): a TrainerDesc object
"""
trainer_desc
.
device_worker_name
=
"HogwildWorker"
class
DownpourSGD
(
DeviceWorker
):
"""
DownpourSGD is a kind of distributed SGD algorithm.
"""
def
__init__
(
self
):
"""
Init.
"""
super
(
DownpourSGD
,
self
).
__init__
()
def
gen_worker_desc
(
self
,
trainer_desc
):
"""
Generator worker desc, which device worker is DownpourWorker.
Args:
trainer_desc(TrainerDesc): a TrainerDesc object
"""
dense_table_set
=
set
()
program_id
=
str
(
id
(
self
.
program_
))
if
self
.
program_
==
None
:
...
...
python/paddle/fluid/incubate/fleet/parameter_server/__init__.py
浏览文件 @
d52586a9
...
...
@@ -127,6 +127,10 @@ class Fleet(object):
init_worker(): will be called by user. When a user knows current process is_server(), he/she
should call init_worker() to initialize global information about worker and connect
worker with pserver.
Args:
programs(Program|list): a Program or a list of Programs
"""
if
not
isinstance
(
programs
,
list
):
programs
=
[
programs
]
...
...
python/paddle/fluid/tests/unittests/test_dataset.py
浏览文件 @
d52586a9
...
...
@@ -21,7 +21,13 @@ import unittest
class
TestDataset
(
unittest
.
TestCase
):
"""
TestCases for Dataset.
"""
def
test_dataset_create
(
self
):
"""
Testcase for dataset create
"""
try
:
dataset
=
fluid
.
DatasetFactory
().
create_dataset
(
"InMemoryDataset"
)
except
:
...
...
@@ -39,6 +45,9 @@ class TestDataset(unittest.TestCase):
self
.
assertTrue
(
True
)
def
test_dataset_config
(
self
):
"""
Testcase for dataset configuration
"""
dataset
=
fluid
.
core
.
Dataset
(
"MultiSlotDataset"
)
dataset
.
set_thread_num
(
12
)
dataset
.
set_filelist
([
"a.txt"
,
"b.txt"
,
"c.txt"
])
...
...
@@ -62,12 +71,15 @@ class TestDataset(unittest.TestCase):
self
.
assertEqual
(
ugi
,
"my_fs_ugi"
)
def
test_in_memory_dataset_run
(
self
):
with
open
(
"test_dataset_a.txt"
,
"w"
)
as
f
:
"""
Testcase for InMemoryDataset from create to run
"""
with
open
(
"test_in_memory_dataset_run_a.txt"
,
"w"
)
as
f
:
data
=
"1 1 2 3 3 4 5 5 5 5 1 1
\n
"
data
+=
"1 2 2 3 4 4 6 6 6 6 1 2
\n
"
data
+=
"1 3 2 3 5 4 7 7 7 7 1 3
\n
"
f
.
write
(
data
)
with
open
(
"test_
dataset
_b.txt"
,
"w"
)
as
f
:
with
open
(
"test_
in_memory_dataset_run
_b.txt"
,
"w"
)
as
f
:
data
=
"1 4 2 3 3 4 5 5 5 5 1 4
\n
"
data
+=
"1 5 2 3 4 4 6 6 6 6 1 5
\n
"
data
+=
"1 6 2 3 5 4 7 7 7 7 1 6
\n
"
...
...
@@ -84,7 +96,8 @@ class TestDataset(unittest.TestCase):
dataset
=
fluid
.
DatasetFactory
().
create_dataset
(
"InMemoryDataset"
)
dataset
.
set_batch_size
(
32
)
dataset
.
set_thread
(
3
)
dataset
.
set_filelist
([
"test_dataset_a.txt"
,
"test_dataset_b.txt"
])
dataset
.
set_filelist
([
"test_in_memory_dataset_run_a.txt"
,
"test_in_memory_dataset_run_b.txt"
])
dataset
.
set_pipe_command
(
"cat"
)
dataset
.
set_use_var
(
slots_vars
)
dataset
.
load_into_memory
()
...
...
@@ -98,16 +111,19 @@ class TestDataset(unittest.TestCase):
except
:
self
.
assertTrue
(
False
)
os
.
remove
(
"./test_
dataset
_a.txt"
)
os
.
remove
(
"./test_
dataset
_b.txt"
)
os
.
remove
(
"./test_
in_memory_dataset_run
_a.txt"
)
os
.
remove
(
"./test_
in_memory_dataset_run
_b.txt"
)
def
test_queue_dataset_run
(
self
):
with
open
(
"test_dataset_a.txt"
,
"w"
)
as
f
:
"""
Testcase for QueueDataset from create to run
"""
with
open
(
"test_queue_dataset_run_a.txt"
,
"w"
)
as
f
:
data
=
"1 1 2 3 3 4 5 5 5 5 1 1
\n
"
data
+=
"1 2 2 3 4 4 6 6 6 6 1 2
\n
"
data
+=
"1 3 2 3 5 4 7 7 7 7 1 3
\n
"
f
.
write
(
data
)
with
open
(
"test_
dataset
_b.txt"
,
"w"
)
as
f
:
with
open
(
"test_
queue_dataset_run
_b.txt"
,
"w"
)
as
f
:
data
=
"1 4 2 3 3 4 5 5 5 5 1 4
\n
"
data
+=
"1 5 2 3 4 4 6 6 6 6 1 5
\n
"
data
+=
"1 6 2 3 5 4 7 7 7 7 1 6
\n
"
...
...
@@ -124,7 +140,8 @@ class TestDataset(unittest.TestCase):
dataset
=
fluid
.
DatasetFactory
().
create_dataset
(
"QueueDataset"
)
dataset
.
set_batch_size
(
32
)
dataset
.
set_thread
(
3
)
dataset
.
set_filelist
([
"test_dataset_a.txt"
,
"test_dataset_b.txt"
])
dataset
.
set_filelist
([
"test_queue_dataset_run_a.txt"
,
"test_queue_dataset_run_b.txt"
])
dataset
.
set_pipe_command
(
"cat"
)
dataset
.
set_use_var
(
slots_vars
)
...
...
@@ -136,8 +153,8 @@ class TestDataset(unittest.TestCase):
except
:
self
.
assertTrue
(
False
)
os
.
remove
(
"./test_
dataset
_a.txt"
)
os
.
remove
(
"./test_
dataset
_b.txt"
)
os
.
remove
(
"./test_
queue_dataset_run
_a.txt"
)
os
.
remove
(
"./test_
queue_dataset_run
_b.txt"
)
if
__name__
==
'__main__'
:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录