Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
f65a61a2
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
f65a61a2
编写于
9月 20, 2022
作者:
Z
zyfncg
提交者:
GitHub
9月 20, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
clear extra attrs of condition op in opmaker (#46150)
上级
d13a4a25
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
50 addition
and
51 deletion
+50
-51
paddle/fluid/operators/controlflow/conditional_block_op.h
paddle/fluid/operators/controlflow/conditional_block_op.h
+0
-5
paddle/fluid/operators/controlflow/while_op.cc
paddle/fluid/operators/controlflow/while_op.cc
+0
-5
paddle/phi/api/yaml/op_compat.yaml
paddle/phi/api/yaml/op_compat.yaml
+50
-41
未找到文件。
paddle/fluid/operators/controlflow/conditional_block_op.h
浏览文件 @
f65a61a2
...
@@ -119,11 +119,6 @@ class ConditionalBlockOpProtoMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -119,11 +119,6 @@ class ConditionalBlockOpProtoMaker : public framework::OpProtoAndCheckerMaker {
"The conditional variable (Cond) is used as scalar "
"The conditional variable (Cond) is used as scalar "
"condition."
)
"condition."
)
.
SetDefault
(
false
);
.
SetDefault
(
false
);
AddAttr
<
std
::
vector
<
std
::
string
>>
(
ConditionalOp
::
kSkipEagerDeletionVars
,
"Vars that would not be deleted when "
"garbage collection strategy enables"
)
.
SetDefault
(
std
::
vector
<
std
::
string
>
())
.
AsExtra
();
AddComment
(
R"DOC(Conditional block operator
AddComment
(
R"DOC(Conditional block operator
If `is_scalar_condition` is True, the conditional variable (Cond) is a scalar,
If `is_scalar_condition` is True, the conditional variable (Cond) is a scalar,
...
...
paddle/fluid/operators/controlflow/while_op.cc
浏览文件 @
f65a61a2
...
@@ -221,11 +221,6 @@ class WhileOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -221,11 +221,6 @@ class WhileOpMaker : public framework::OpProtoAndCheckerMaker {
"(bool, default false) Set to true for inference only, false "
"(bool, default false) Set to true for inference only, false "
"for training. Some layers may run faster when this is true."
)
"for training. Some layers may run faster when this is true."
)
.
SetDefault
(
false
);
.
SetDefault
(
false
);
AddAttr
<
std
::
vector
<
std
::
string
>>
(
kSkipEagerDeletionVars
,
"Vars that would skip eager deletion."
"Users should not set this manually."
)
.
SetDefault
(
std
::
vector
<
std
::
string
>
())
.
AsExtra
();
AddComment
(
R"DOC(
AddComment
(
R"DOC(
)DOC"
);
)DOC"
);
}
}
...
...
paddle/phi/api/yaml/op_compat.yaml
浏览文件 @
f65a61a2
...
@@ -93,6 +93,11 @@
...
@@ -93,6 +93,11 @@
extra
:
extra
:
attrs
:
[
bool use_mkldnn = false
,
bool use_quantizer = false
,
str mkldnn_data_type = "float32"
]
attrs
:
[
bool use_mkldnn = false
,
bool use_quantizer = false
,
str mkldnn_data_type = "float32"
]
-
op
:
conditional_block
backward
:
conditional_block_grad
extra
:
attrs
:
[
'
str[]
skip_eager_deletion_vars
=
{}'
]
-
op
:
conv2d
-
op
:
conv2d
backward
:
conv2d_grad
backward
:
conv2d_grad
extra
:
extra
:
...
@@ -249,6 +254,11 @@
...
@@ -249,6 +254,11 @@
extra
:
extra
:
attrs
:
[
bool use_mkldnn = false
,
bool use_cudnn = false
]
attrs
:
[
bool use_mkldnn = false
,
bool use_cudnn = false
]
-
op
:
expand (expand_v2)
backward
:
expand_grad (expand_v2_grad)
extra
:
attrs
:
[
bool use_mkldnn = false
,
str mkldnn_data_type = "float32"
]
-
op
:
expm1
-
op
:
expm1
backward
:
expm1_grad
backward
:
expm1_grad
extra
:
extra
:
...
@@ -293,6 +303,15 @@
...
@@ -293,6 +303,15 @@
extra
:
extra
:
attrs
:
[
bool use_mkldnn = false
]
attrs
:
[
bool use_mkldnn = false
]
-
op
:
full (fill_constant)
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
gather
backward
:
gather_grad
extra
:
attrs
:
[
bool overwrite = true
]
-
op
:
gelu
-
op
:
gelu
backward
:
gelu_grad
backward
:
gelu_grad
extra
:
extra
:
...
@@ -392,6 +411,12 @@
...
@@ -392,6 +411,12 @@
str mkldnn_data_type = "float32"
,
'
int[]
fused_reshape_X
=
{}'
,
'
int[]
fused_reshape_Y
=
{}'
,
str mkldnn_data_type = "float32"
,
'
int[]
fused_reshape_X
=
{}'
,
'
int[]
fused_reshape_Y
=
{}'
,
'
int[]
fused_transpose_X
=
{}'
,
'
int[]
fused_transpose_Y
=
{}'
]
'
int[]
fused_transpose_X
=
{}'
,
'
int[]
fused_transpose_Y
=
{}'
]
-
op
:
matmul_with_flatten (mul)
backward
:
matmul_with_flatten_grad (mul_grad)
extra
:
attrs
:
[
bool use_mkldnn = false
,
float scale_x = 1.0f
,
'
float[]
scale_y
=
{1.0f}'
,
float scale_out = 1.0f
,
bool force_fp32_output = false
]
-
op
:
maximum (elementwise_max)
-
op
:
maximum (elementwise_max)
backward
:
maximum_grad (elementwise_max_grad)
backward
:
maximum_grad (elementwise_max_grad)
extra
:
extra
:
...
@@ -447,6 +472,17 @@
...
@@ -447,6 +472,17 @@
outputs
:
outputs
:
out
:
Out
out
:
Out
-
op
:
pool2d
backward
:
pool2d_grad
extra
:
attrs
:
[
bool use_mkldnn = false
,
bool use_quantizer = false
,
str mkldnn_data_type = "float32"
,
bool is_test = false
]
-
op
:
pool3d
backward
:
pool3d_grad
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
prelu
-
op
:
prelu
backward
:
prelu_grad
backward
:
prelu_grad
extra
:
extra
:
...
@@ -619,6 +655,11 @@
...
@@ -619,6 +655,11 @@
extra
:
extra
:
attrs
:
[
bool use_mkldnn = false
]
attrs
:
[
bool use_mkldnn = false
]
-
op
:
stack
backward
:
stack_grad
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
subtract (elementwise_sub)
-
op
:
subtract (elementwise_sub)
backward
:
subtract_grad (elementwise_sub_grad)
backward
:
subtract_grad (elementwise_sub_grad)
extra
:
extra
:
...
@@ -656,6 +697,12 @@
...
@@ -656,6 +697,12 @@
outputs
:
outputs
:
out
:
Out
out
:
Out
-
op
:
transpose (transpose2)
backward
:
transpose_grad (transpose2_grad)
extra
:
attrs
:
[
bool use_mkldnn = false
,
str data_format = "AnyLayout"
,
bool use_quantizer = false
,
str mkldnn_data_type = "float32"
]
-
op
:
trilinear_interp (trilinear_interp_v2)
-
op
:
trilinear_interp (trilinear_interp_v2)
backward
:
trilinear_interp_grad (trilinear_interp_v2_grad)
backward
:
trilinear_interp_grad (trilinear_interp_v2_grad)
extra
:
extra
:
...
@@ -667,45 +714,7 @@
...
@@ -667,45 +714,7 @@
outputs
:
outputs
:
out
:
Out
out
:
Out
-
op
:
expand (expand_v2)
-
op
:
while
backward
:
expand_grad (expand_v2_grad)
backward
:
while_grad
extra
:
attrs
:
[
bool use_mkldnn = false
,
str mkldnn_data_type = "float32"
]
-
op
:
full (fill_constant)
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
gather
backward
:
gather_grad
extra
:
attrs
:
[
bool overwrite = true
]
-
op
:
matmul_with_flatten (mul)
backward
:
matmul_with_flatten_grad (mul_grad)
extra
:
attrs
:
[
bool use_mkldnn = false
,
float scale_x = 1.0f
,
'
float[]
scale_y
=
{1.0f}'
,
float scale_out = 1.0f
,
bool force_fp32_output = false
]
-
op
:
pool2d
backward
:
pool2d_grad
extra
:
extra
:
attrs
:
[
bool use_mkldnn = false
,
bool use_quantizer = false
,
attrs
:
[
'
str[]
skip_eager_deletion_vars
=
{}'
]
str mkldnn_data_type = "float32"
,
bool is_test = false
]
-
op
:
pool3d
backward
:
pool3d_grad
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
stack
backward
:
stack_grad
extra
:
attrs
:
[
bool use_mkldnn = false
]
-
op
:
transpose (transpose2)
backward
:
transpose_grad (transpose2_grad)
extra
:
attrs
:
[
bool use_mkldnn = false
,
str data_format = "AnyLayout"
,
bool use_quantizer = false
,
str mkldnn_data_type = "float32"
]
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录