Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
s920243400
PaddleDetection
提交
b97d61ad
P
PaddleDetection
项目概览
s920243400
/
PaddleDetection
与 Fork 源项目一致
Fork自
PaddlePaddle / PaddleDetection
通知
2
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
b97d61ad
编写于
2月 24, 2018
作者:
Q
qijun
浏览文件
操作
浏览文件
下载
差异文件
merge baidu/develop
上级
50cf103e
d4dabe3e
变更
13
隐藏空白更改
内联
并排
Showing
13 changed file
with
180 addition
and
60 deletion
+180
-60
benchmark/cluster/vgg16/README.md
benchmark/cluster/vgg16/README.md
+5
-4
paddle/fluid/framework/block_desc.cc
paddle/fluid/framework/block_desc.cc
+43
-9
paddle/fluid/framework/block_desc.h
paddle/fluid/framework/block_desc.h
+7
-1
paddle/fluid/framework/channel.h
paddle/fluid/framework/channel.h
+2
-3
paddle/fluid/framework/channel_test.cc
paddle/fluid/framework/channel_test.cc
+34
-0
paddle/fluid/framework/framework.proto
paddle/fluid/framework/framework.proto
+1
-0
paddle/fluid/framework/program_desc.h
paddle/fluid/framework/program_desc.h
+7
-1
paddle/fluid/operators/while_op.cc
paddle/fluid/operators/while_op.cc
+15
-10
paddle/fluid/pybind/protobuf.cc
paddle/fluid/pybind/protobuf.cc
+2
-0
python/paddle/v2/fluid/backward.py
python/paddle/v2/fluid/backward.py
+4
-1
python/paddle/v2/fluid/framework.py
python/paddle/v2/fluid/framework.py
+49
-20
python/paddle/v2/fluid/layers/nn.py
python/paddle/v2/fluid/layers/nn.py
+1
-1
python/paddle/v2/fluid/memory_optimization_transpiler.py
python/paddle/v2/fluid/memory_optimization_transpiler.py
+10
-10
未找到文件。
benchmark/cluster/vgg16/README.md
浏览文件 @
b97d61ad
...
...
@@ -8,23 +8,24 @@
-
cpu MHz : 2101.000
-
cache size : 20480 KB
### Blas settings
Setting environment variable:
`MKL_NUM_THREADS=1`
.
### Single Node Single Thread
-
PServer Count: 10
-
Trainer Count: 20
-
Metrics: samples / sec
| Batch Size | 32 | 64 | 128 | 256 |
| -- | -- | -- | -- | -- |
| PaddlePaddle Fluid | 15.44 | 16.32 | 16.74 | 16.79 |
| PaddlePaddle v2 | 15.97 | 17.04 | 17.60 | 17.83 |
| TensorFlow |
- | - | - | -
|
| TensorFlow |
9.09 | 9.10 | 9.24 | 8.66
|
### Different Batch Size
-
PServer Count: 10
-
Trainer Count: 20
-
Per trainer CPU Core: 1
-
Metrics: samples / sec
| Batch Size | 32 | 64 | 128 | 256 |
...
...
paddle/fluid/framework/block_desc.cc
浏览文件 @
b97d61ad
...
...
@@ -16,6 +16,8 @@ limitations under the License. */
#include "paddle/fluid/framework/operator.h"
#include "paddle/fluid/framework/program_desc.h"
#include <queue>
namespace
paddle
{
namespace
framework
{
...
...
@@ -64,12 +66,36 @@ VarDesc *BlockDesc::RenameVar(const std::string &old_name,
VarDesc
*
BlockDesc
::
FindVarRecursive
(
const
std
::
string
&
name
)
const
{
if
(
name
==
kEmptyVarName
)
return
nullptr
;
auto
it
=
vars_
.
find
(
name
);
if
(
it
==
vars_
.
end
())
{
return
Parent
()
==
kNoneBlockIndex
?
nullptr
:
ParentBlock
()
->
FindVarRecursive
(
name
);
std
::
queue
<
const
BlockDesc
*>
frontier
;
std
::
unordered_set
<
const
BlockDesc
*>
visited
;
frontier
.
push
(
this
);
while
(
!
frontier
.
empty
())
{
// BFS
auto
cur
=
frontier
.
front
();
frontier
.
pop
();
if
(
visited
.
count
(
cur
)
!=
0
)
{
continue
;
}
auto
var
=
cur
->
FindVar
(
name
);
if
(
var
!=
nullptr
)
{
return
var
;
}
auto
fwd
=
cur
->
ForwardBlock
();
auto
parent
=
cur
->
ParentBlock
();
if
(
fwd
!=
nullptr
)
{
frontier
.
push
(
fwd
);
}
if
(
parent
!=
nullptr
)
{
frontier
.
push
(
parent
);
}
visited
.
insert
(
cur
);
}
return
it
->
second
.
get
();
return
nullptr
;
}
VarDesc
&
BlockDesc
::
FindRecursiveOrCreateVar
(
const
std
::
string
&
name_bytes
)
{
...
...
@@ -155,10 +181,7 @@ void BlockDesc::Flush() {
}
BlockDesc
*
BlockDesc
::
ParentBlock
()
const
{
if
(
this
->
desc_
->
parent_idx
()
==
kNoneBlockIndex
)
{
return
nullptr
;
}
return
prog_
->
MutableBlock
(
static_cast
<
size_t
>
(
this
->
desc_
->
parent_idx
()));
return
prog_
->
MutableBlock
(
static_cast
<
size_t
>
(
desc_
->
parent_idx
()));
}
proto
::
BlockDesc
*
BlockDesc
::
Proto
()
{
...
...
@@ -205,5 +228,16 @@ void BlockDesc::ClearPBVars() {
}
}
void
BlockDesc
::
SetForwardBlockID
(
int32_t
forward_block_id
)
{
PADDLE_ENFORCE
(
!
desc_
->
has_forward_block_idx
(),
"Parent block ID has been set to %d. Cannot set to %d"
,
desc_
->
forward_block_idx
(),
forward_block_id
);
desc_
->
set_forward_block_idx
(
forward_block_id
);
}
BlockDesc
*
BlockDesc
::
ForwardBlock
()
const
{
return
prog_
->
MutableBlock
(
static_cast
<
size_t
>
(
desc_
->
forward_block_idx
()));
}
}
// namespace framework
}
// namespace paddle
paddle/fluid/framework/block_desc.h
浏览文件 @
b97d61ad
...
...
@@ -49,6 +49,8 @@ class BlockDesc {
int32_t
Parent
()
const
{
return
desc_
->
parent_idx
();
}
int32_t
ForwardBlockID
()
const
{
return
desc_
->
forward_block_idx
();
}
VarDesc
*
Var
(
const
std
::
string
&
name_bytes
);
VarDesc
*
FindVar
(
const
std
::
string
&
name_bytes
)
const
;
...
...
@@ -75,6 +77,10 @@ class BlockDesc {
BlockDesc
*
ParentBlock
()
const
;
BlockDesc
*
ForwardBlock
()
const
;
void
SetForwardBlockID
(
int32_t
forward_block_id
);
OpDesc
*
AppendOp
();
void
AppendAllocatedOp
(
std
::
unique_ptr
<
OpDesc
>
&&
op_desc
);
...
...
@@ -93,7 +99,7 @@ class BlockDesc {
proto
::
BlockDesc
*
Proto
();
ProgramDesc
*
Program
()
{
return
this
->
prog_
;
}
ProgramDesc
*
Program
()
const
{
return
this
->
prog_
;
}
private:
void
ClearPBOps
();
...
...
paddle/fluid/framework/channel.h
浏览文件 @
b97d61ad
...
...
@@ -100,8 +100,7 @@ class ChannelHolder {
virtual
~
Placeholder
()
{}
virtual
const
std
::
type_index
Type
()
const
=
0
;
virtual
void
*
Ptr
()
const
=
0
;
virtual
void
Close
()
const
=
0
;
std
::
type_info
type_
;
virtual
void
Close
()
=
0
;
};
template
<
typename
T
>
...
...
@@ -116,7 +115,7 @@ class ChannelHolder {
if
(
channel_
)
channel_
->
Close
();
}
std
::
unique_ptr
<
Channel
<
T
>
*
>
channel_
;
std
::
unique_ptr
<
Channel
<
T
>>
channel_
;
const
std
::
type_index
type_
;
};
...
...
paddle/fluid/framework/channel_test.cc
浏览文件 @
b97d61ad
...
...
@@ -20,6 +20,7 @@ limitations under the License. */
#include "gtest/gtest.h"
using
paddle
::
framework
::
Channel
;
using
paddle
::
framework
::
ChannelHolder
;
using
paddle
::
framework
::
MakeChannel
;
using
paddle
::
framework
::
CloseChannel
;
using
paddle
::
framework
::
details
::
Buffered
;
...
...
@@ -508,3 +509,36 @@ TEST(Channel, UnbufferedChannelDestroyUnblocksSendersTest) {
auto
ch
=
MakeChannel
<
int
>
(
0
);
ChannelDestroyUnblockSenders
(
ch
);
}
void
ChannelHolderSendReceive
(
ChannelHolder
*
ch
)
{
unsigned
sum_send
=
0
;
std
::
thread
t
([
&
]()
{
for
(
int
i
=
0
;
i
<
5
;
i
++
)
{
EXPECT_EQ
(
ch
->
Send
(
&
i
),
true
);
sum_send
+=
i
;
}
});
for
(
int
i
=
0
;
i
<
5
;
i
++
)
{
int
recv
;
EXPECT_EQ
(
ch
->
Receive
(
&
recv
),
true
);
EXPECT_EQ
(
recv
,
i
);
}
ch
->
close
();
t
.
join
();
EXPECT_EQ
(
sum_send
,
10U
);
}
TEST
(
ChannelHolder
,
ChannelHolderBufferedSendReceiveTest
)
{
ChannelHolder
*
ch
=
new
ChannelHolder
();
ch
->
Reset
<
int
>
(
10
);
ChannelHolderSendReceive
(
ch
);
delete
ch
;
}
TEST
(
ChannelHolder
,
ChannelHolderUnBufferedSendReceiveTest
)
{
ChannelHolder
*
ch
=
new
ChannelHolder
();
ch
->
Reset
<
int
>
(
0
);
ChannelHolderSendReceive
(
ch
);
delete
ch
;
}
paddle/fluid/framework/framework.proto
浏览文件 @
b97d61ad
...
...
@@ -158,6 +158,7 @@ message BlockDesc {
required
int32
parent_idx
=
2
;
repeated
VarDesc
vars
=
3
;
repeated
OpDesc
ops
=
4
;
optional
int32
forward_block_idx
=
5
[
default
=
-
1
];
}
// Please refer to
...
...
paddle/fluid/framework/program_desc.h
浏览文件 @
b97d61ad
...
...
@@ -38,7 +38,13 @@ class ProgramDesc {
BlockDesc
*
AppendBlock
(
const
BlockDesc
&
parent
);
BlockDesc
*
MutableBlock
(
size_t
idx
)
{
return
blocks_
[
idx
].
get
();
}
BlockDesc
*
MutableBlock
(
size_t
idx
)
{
if
(
idx
==
static_cast
<
size_t
>
(
kNoneBlockIndex
))
{
return
nullptr
;
}
else
{
return
blocks_
[
idx
].
get
();
}
}
const
BlockDesc
&
Block
(
size_t
idx
)
const
{
return
*
blocks_
[
idx
];
}
...
...
paddle/fluid/operators/while_op.cc
浏览文件 @
b97d61ad
...
...
@@ -231,7 +231,8 @@ class WhileGradOpDescMaker : public framework::SingleGradOpDescMaker {
while_grad
->
SetInput
(
kStepScopes
,
Output
(
kStepScopes
));
auto
*
grad_block
=
this
->
grad_block_
[
0
];
auto
*
fwd_block
=
grad_block
->
ParentBlock
();
auto
*
fwd_block
=
grad_block
->
ForwardBlock
();
auto
*
parent_block
=
grad_block
->
ParentBlock
();
// Not all of IGs will be generated by inner gradient operators of while op.
// Ignore IGs that is not generated by the inside block.
...
...
@@ -260,33 +261,37 @@ class WhileGradOpDescMaker : public framework::SingleGradOpDescMaker {
for
(
auto
&
o
:
Output
(
kOutputs
))
{
block_ins
.
insert
(
o
);
}
std
::
unordered_set
<
std
::
string
>
extra_input
s
;
std
::
unordered_set
<
std
::
string
>
output_grad
s
;
for
(
const
auto
*
op
:
grad_block
->
AllOps
())
{
for
(
auto
&
input_name
:
op
->
InputArgumentNames
())
{
// If the input of Op has been recorded or is generated by the forward
// block, do not make it as input again.
// The input is located in I/O or other op's outputs or the variable is
// located in grad_block's parents
if
(
block_ins
.
find
(
input_name
)
!=
block_ins
.
end
()
||
fwd_block
->
FindVar
(
input_name
)
!=
nullptr
)
{
(
fwd_block
->
FindVarRecursive
(
input_name
)
!=
nullptr
||
parent_block
->
FindVarRecursive
(
input_name
)
!=
nullptr
))
{
continue
;
}
extra_input
s
.
insert
(
input_name
);
output_grad
s
.
insert
(
input_name
);
}
for
(
auto
&
output_name
:
op
->
OutputArgumentNames
())
{
block_ins
.
insert
(
output_name
);
}
}
std
::
vector
<
std
::
string
>
extra_input
s_list
;
extra_inputs_list
.
resize
(
extra_input
s
.
size
());
std
::
copy
(
extra_inputs
.
begin
(),
extra_input
s
.
end
(),
extra_input
s_list
.
begin
());
while_grad
->
SetInput
(
framework
::
GradVarName
(
kOutputs
),
extra_input
s_list
);
std
::
vector
<
std
::
string
>
output_grad
s_list
;
output_grads_list
.
resize
(
output_grad
s
.
size
());
std
::
copy
(
output_grads
.
begin
(),
output_grad
s
.
end
(),
output_grad
s_list
.
begin
());
while_grad
->
SetInput
(
framework
::
GradVarName
(
kOutputs
),
output_grad
s_list
);
while_grad
->
SetAttrMap
(
this
->
Attrs
());
while_grad
->
SetBlockAttr
(
kStepBlock
,
*
grad_block
);
// record the original output gradient names, since the gradient name of
// while operator could be renamed.
while_grad
->
SetAttr
(
"original_output_grad"
,
extra_input
s_list
);
while_grad
->
SetAttr
(
"original_output_grad"
,
output_grad
s_list
);
return
std
::
unique_ptr
<
framework
::
OpDesc
>
(
while_grad
);
}
...
...
paddle/fluid/pybind/protobuf.cc
浏览文件 @
b97d61ad
...
...
@@ -155,6 +155,8 @@ void BindBlockDesc(py::module &m) {
py
::
class_
<
BlockDesc
>
(
m
,
"BlockDesc"
,
""
)
.
def_property_readonly
(
"id"
,
&
BlockDesc
::
ID
)
.
def_property_readonly
(
"parent"
,
&
BlockDesc
::
Parent
)
.
def
(
"get_forward_block_idx"
,
&
BlockDesc
::
ForwardBlockID
)
.
def
(
"set_forward_block_idx"
,
&
BlockDesc
::
SetForwardBlockID
)
.
def
(
"append_op"
,
&
BlockDesc
::
AppendOp
,
py
::
return_value_policy
::
reference
)
.
def
(
"prepend_op"
,
&
BlockDesc
::
PrependOp
,
...
...
python/paddle/v2/fluid/backward.py
浏览文件 @
b97d61ad
...
...
@@ -298,7 +298,8 @@ def _append_backward_ops_(block,
# If the op has its own sub-block, deal with the sub-block first
if
op
.
has_attr
(
"sub_block"
):
sub_block
=
program
.
block
(
op
.
block_attr
(
"sub_block"
))
grad_sub_block
=
program
.
create_block
(
parent_idx
=
sub_block
.
idx
)
grad_sub_block
=
program
.
create_block
()
grad_sub_block
.
set_forward_block_idx
(
sub_block
.
idx
)
cb
=
_callback_lookup_
(
op
)
if
cb
is
not
None
:
if
callbacks
is
None
:
...
...
@@ -310,6 +311,8 @@ def _append_backward_ops_(block,
else
:
_append_backward_ops_
(
sub_block
,
sub_block
.
ops
,
grad_sub_block
,
no_grad_dict
,
grad_to_var
,
callbacks
)
program
.
rollback
()
grad_sub_block_list
.
append
(
grad_sub_block
.
desc
)
# Getting op's corresponding grad_op
...
...
python/paddle/v2/fluid/framework.py
浏览文件 @
b97d61ad
...
...
@@ -152,7 +152,7 @@ class Variable(object):
shape(tuple|list|None): The shape of variable. -1 means the batch size.
Some kinds of variable do not contain shape, just set it to None.
dtype(np.dtype|core.VarDesc.VarType|str): The data type of variable.
lod_level(int): The level of lod tensor. 0 means
there
is not a time
lod_level(int): The level of lod tensor. 0 means
it
is not a time
series data.
persistable(bool): True if the variable should be saved as check point.
Defaults to False.
...
...
@@ -346,7 +346,7 @@ class OpProtoHolder(object):
def
__init__
(
self
):
assert
not
hasattr
(
self
.
__class__
,
'_instance'
),
'Please use `instance()` to get OpProtoHolder o
p
ject!'
'_instance'
),
'Please use `instance()` to get OpProtoHolder o
b
ject!'
op_protos
=
get_all_op_protos
()
self
.
op_proto_map
=
{}
for
proto
in
op_protos
:
...
...
@@ -368,8 +368,8 @@ class OpProtoHolder(object):
class
Operator
(
object
):
"""
Python Operator class. The operator represents the build in instructs in a
Block. Users can use the build in instructs to describe their neural
Python Operator class. The operator represents the build in instruct
ion
s in a
Block. Users can use the build in instruct
ion
s to describe their neural
network.
"""
...
...
@@ -478,7 +478,7 @@ class Operator(object):
raise
TypeError
(
"'attrs' should be a dict."
)
for
attr
in
proto
.
attrs
:
attr_name
=
attr
.
name
if
(
not
attr_name
in
attrs
)
or
(
attrs
[
attr_name
]
is
None
):
if
(
attr_name
not
in
attrs
)
or
(
attrs
[
attr_name
]
is
None
):
continue
if
isinstance
(
attrs
[
attr_name
],
Block
):
self
.
desc
.
set_block_attr
(
attr_name
,
attrs
[
attr_name
].
desc
)
...
...
@@ -696,6 +696,13 @@ class Block(object):
def
parent_idx
(
self
):
return
self
.
desc
.
parent
@
property
def
forward_block_idx
(
self
):
return
self
.
desc
.
get_forward_block_idx
()
def
set_forward_block_idx
(
self
,
idx
):
self
.
desc
.
set_forward_block_idx
(
idx
)
@
property
def
idx
(
self
):
return
self
.
desc
.
id
...
...
@@ -709,15 +716,32 @@ class Block(object):
return
v
def
var_recursive
(
self
,
name
):
if
self
.
has_var
(
name
):
return
self
.
var
(
name
)
else
:
if
self
.
idx
==
0
:
raise
ValueError
(
"var %s is not in block(%d) nor its parents."
%
name
,
self
.
idx
)
else
:
parent_block
=
self
.
program
.
block
(
self
.
parent_idx
)
return
parent_block
.
var_recursive
(
name
)
frontier
=
list
()
visited
=
set
()
frontier
.
append
(
self
)
prog
=
self
.
program
while
len
(
frontier
)
!=
0
:
# BFS
cur
=
frontier
[
0
]
frontier
=
frontier
[
1
:]
if
id
(
cur
)
in
visited
:
continue
if
cur
.
has_var
(
name
):
return
cur
.
var
(
name
)
if
cur
.
parent_idx
!=
-
1
:
frontier
.
append
(
prog
.
block
(
cur
.
parent_idx
))
if
cur
.
forward_block_idx
!=
-
1
:
frontier
.
append
(
prog
.
block
(
cur
.
forward_block_idx
))
visited
.
add
(
id
(
cur
))
raise
ValueError
(
"Var {0} is not found recursively"
.
format
(
name
))
def
all_parameters
(
self
):
return
list
(
self
.
iter_parameters
())
...
...
@@ -727,7 +751,7 @@ class Block(object):
if
isinstance
(
item
[
1
],
Parameter
))
def
create_var
(
self
,
*
args
,
**
kwargs
):
var
=
Variable
(
self
,
*
args
,
**
kwargs
)
var
=
Variable
(
block
=
self
,
*
args
,
**
kwargs
)
if
'initializer'
in
kwargs
:
kwargs
[
'initializer'
](
var
,
self
)
return
var
...
...
@@ -798,13 +822,13 @@ class Block(object):
def
append_op
(
self
,
*
args
,
**
kwargs
):
op_desc
=
self
.
desc
.
append_op
()
op
=
Operator
(
self
,
op_desc
,
*
args
,
**
kwargs
)
op
=
Operator
(
block
=
self
,
desc
=
op_desc
,
*
args
,
**
kwargs
)
self
.
ops
.
append
(
op
)
return
op
def
delete_ops
(
self
,
ops
):
# remove from cpp
# FIXME(typhoonzero): remove only the first occur
acy
.
# FIXME(typhoonzero): remove only the first occur
rence
.
try
:
start
=
list
(
self
.
ops
).
index
(
ops
[
0
])
end
=
list
(
self
.
ops
).
index
(
ops
[
-
1
])
...
...
@@ -822,6 +846,11 @@ class Block(object):
return
op
def
sync_with_cpp
(
self
):
"""
Sync with the desc on the c++ end.
This method is used to synchronize the c++ desc instance generated by backward.
"""
# sync variables from cpp
for
var
in
self
.
desc
.
all_vars
():
if
not
self
.
has_var
(
var
.
name
()):
...
...
@@ -867,9 +896,9 @@ class Block(object):
def
copy_param_info_from
(
self
,
other
):
"""
Copy the information of parameters from other block
Copy the information of parameters from
the
other block
Args:
other(Block): other block
other(Block):
the
other block
Returns:
None
...
...
@@ -1215,6 +1244,6 @@ def get_var(name, program=None):
if
program
is
None
:
program
=
default_main_program
()
assert
isinstance
(
name
,
str
)
assert
isinstance
(
name
,
Program
)
assert
isinstance
(
program
,
Program
)
return
program
.
global_block
().
var
(
name
)
python/paddle/v2/fluid/layers/nn.py
浏览文件 @
b97d61ad
...
...
@@ -104,7 +104,7 @@ def fc(input,
* :math:`X_i`: The input tensor.
* :math:`W`: The weights created by this layer.
* :math:`b`: The bias parameter created by this layer (if needed).
* :math:`Act`: The activation funtion.
* :math:`Act`: The activation fun
c
tion.
* :math:`Out`: The output tensor.
Args:
...
...
python/paddle/v2/fluid/memory_optimization_transpiler.py
浏览文件 @
b97d61ad
...
...
@@ -220,15 +220,15 @@ def _process_sub_block_pair(pdesc, sub_block_pair):
# Find fwd_op/bwd_op block pair
for
grad_id
in
grad_sub_block_ids
:
parent_id
=
pdesc
.
block
(
grad_id
).
parent
if
parent
_id
in
sub_block_ids
:
sub_block_id_pair
.
append
((
parent
_id
,
grad_id
))
sub_block_ids
.
remove
(
parent
_id
)
fwd_id
=
pdesc
.
block
(
grad_id
).
get_forward_block_idx
()
if
fwd
_id
in
sub_block_ids
:
sub_block_id_pair
.
append
((
fwd
_id
,
grad_id
))
sub_block_ids
.
remove
(
fwd
_id
)
# Get fwd_op/bwd_op block ops
for
parent
_id
,
grad_id
in
sub_block_id_pair
:
for
fwd
_id
,
grad_id
in
sub_block_id_pair
:
sub_block_ops
=
[]
sub_block
=
pdesc
.
block
(
parent
_id
)
sub_block
=
pdesc
.
block
(
fwd
_id
)
block_op_size
=
sub_block
.
op_size
()
for
i
in
range
(
block_op_size
):
sub_block_ops
.
append
(
sub_block
.
op
(
i
))
...
...
@@ -239,19 +239,19 @@ def _process_sub_block_pair(pdesc, sub_block_pair):
sub_block_ops
.
append
(
grad_sub_block
.
op
(
i
))
sub_op_output
=
set
()
sub_op_output
.
update
(
sub_op_dict
[
parent
_id
].
output_arg_names
())
sub_op_output
.
update
(
sub_op_dict
[
fwd
_id
].
output_arg_names
())
sub_op_output
.
update
(
sub_op_dict
[
grad_id
].
output_arg_names
())
ops_list
.
append
((
sub_block_ops
,
block_op_size
,
sub_op_output
))
# Process rest fwd_op block ops
for
parent
_id
in
sub_block_ids
:
for
fwd
_id
in
sub_block_ids
:
sub_block_ops
=
[]
sub_block
=
pdesc
.
block
(
parent
_id
)
sub_block
=
pdesc
.
block
(
fwd
_id
)
sub_block_op_size
=
sub_block
.
op_size
()
for
i
in
range
(
sub_block_op_size
):
sub_block_ops
.
append
(
sub_block
.
op
(
i
))
sub_op_output
=
set
()
sub_op_output
.
update
(
sub_op_dict
[
parent
_id
].
output_arg_names
())
sub_op_output
.
update
(
sub_op_dict
[
fwd
_id
].
output_arg_names
())
ops_list
.
append
((
sub_block_ops
,
sub_block_op_size
,
sub_op_output
))
return
ops_list
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录