Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
e33b4112
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
e33b4112
编写于
9月 28, 2017
作者:
G
guosheng
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Adapt reduce_op according to up-to-date dev
上级
be58c632
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
21 addition
and
20 deletion
+21
-20
paddle/operators/reduce_op.cc
paddle/operators/reduce_op.cc
+21
-20
未找到文件。
paddle/operators/reduce_op.cc
浏览文件 @
e33b4112
...
@@ -24,20 +24,20 @@ class ReduceOp : public framework::OperatorWithKernel {
...
@@ -24,20 +24,20 @@ class ReduceOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
protected:
protected:
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContextBase
*
ctx
)
const
override
{
PADDLE_ENFORCE
_NOT_NULL
(
ctx
.
InputVar
(
"X"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
"Input(X) of ReduceOp should not be null."
);
"Input(X) of ReduceOp should not be null."
);
PADDLE_ENFORCE
_NOT_NULL
(
ctx
.
OutputVar
(
"Out"
),
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
"Output(Out) of ReduceOp should not be null."
);
"Output(Out) of ReduceOp should not be null."
);
auto
x_dims
=
ctx
.
Input
<
Tensor
>
(
"X"
)
->
dims
(
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
x_rank
=
x_dims
.
size
();
auto
x_rank
=
x_dims
.
size
();
PADDLE_ENFORCE_LE
(
x_rank
,
6
,
"Tensors with rank at most 6 are supported."
);
PADDLE_ENFORCE_LE
(
x_rank
,
6
,
"Tensors with rank at most 6 are supported."
);
int
dim
=
ctx
.
Attr
<
int
>
(
"dim"
);
int
dim
=
ctx
->
Attrs
().
Get
<
int
>
(
"dim"
);
if
(
dim
<
0
)
dim
=
x_rank
+
dim
;
if
(
dim
<
0
)
dim
=
x_rank
+
dim
;
PADDLE_ENFORCE_LT
(
PADDLE_ENFORCE_LT
(
dim
,
x_rank
,
dim
,
x_rank
,
"The dim should be in the range [-rank(input), rank(input))."
);
"The dim should be in the range [-rank(input), rank(input))."
);
bool
keep_dim
=
ctx
.
Attr
<
bool
>
(
"keep_dim"
);
bool
keep_dim
=
ctx
->
Attrs
().
Get
<
bool
>
(
"keep_dim"
);
auto
dims_vector
=
vectorize
(
x_dims
);
auto
dims_vector
=
vectorize
(
x_dims
);
if
(
keep_dim
||
x_rank
==
1
)
{
if
(
keep_dim
||
x_rank
==
1
)
{
dims_vector
[
dim
]
=
1
;
dims_vector
[
dim
]
=
1
;
...
@@ -45,10 +45,10 @@ class ReduceOp : public framework::OperatorWithKernel {
...
@@ -45,10 +45,10 @@ class ReduceOp : public framework::OperatorWithKernel {
dims_vector
.
erase
(
dims_vector
.
begin
()
+
dim
);
dims_vector
.
erase
(
dims_vector
.
begin
()
+
dim
);
}
}
auto
out_dims
=
framework
::
make_ddim
(
dims_vector
);
auto
out_dims
=
framework
::
make_ddim
(
dims_vector
);
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
)
->
Resize
(
out_dims
);
ctx
->
SetOutputDim
(
"Out"
,
out_dims
);
if
(
dim
!=
0
)
{
if
(
dim
!=
0
)
{
// Only pass LoD when not reducing on the first dim
// Only pass LoD when not reducing on the first dim
.
ctx
.
ShareLoD
(
"X"
,
/*->*/
"Out"
);
ctx
->
ShareLoD
(
"X"
,
/*->*/
"Out"
);
}
}
}
}
};
};
...
@@ -58,21 +58,22 @@ class ReduceGradOp : public framework::OperatorWithKernel {
...
@@ -58,21 +58,22 @@ class ReduceGradOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
protected:
protected:
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContextBase
*
ctx
)
const
override
{
PADDLE_ENFORCE
_NOT_NULL
(
ctx
.
InputVar
(
"X"
),
"Input(X) should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
"Input(X) should not be null."
);
PADDLE_ENFORCE
_NOT_NULL
(
ctx
.
InputVar
(
framework
::
GradVarName
(
"Out"
)),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
"Input(Out@GRAD) should not be null."
);
"Input(Out@GRAD) should not be null."
);
auto
x_dims
=
ctx
.
Input
<
Tensor
>
(
"X"
)
->
dims
(
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
x_rank
=
x_dims
.
size
();
auto
x_rank
=
x_dims
.
size
();
PADDLE_ENFORCE_LE
(
x_rank
,
6
,
"Tensors with rank at most 6 are supported."
);
PADDLE_ENFORCE_LE
(
x_rank
,
6
,
"Tensors with rank at most 6 are supported."
);
int
dim
=
ctx
.
Attr
<
int
>
(
"dim"
);
int
dim
=
ctx
->
Attrs
().
Get
<
int
>
(
"dim"
);
if
(
dim
<
0
)
dim
=
x_rank
+
dim
;
if
(
dim
<
0
)
dim
=
x_rank
+
dim
;
PADDLE_ENFORCE_LT
(
PADDLE_ENFORCE_LT
(
dim
,
x_rank
,
dim
,
x_rank
,
"The dim should be in the range [-rank(input), rank(input))."
);
"The dim should be in the range [-rank(input), rank(input))."
);
auto
*
x_grad
=
auto
x_grad_name
=
framework
::
GradVarName
(
"X"
);
ctx
.
Output
<
framework
::
LoDTensor
>
(
framework
::
GradVarName
(
"X"
));
if
(
ctx
->
HasOutput
(
x_grad_name
))
{
if
(
x_grad
)
x_grad
->
Resize
(
x_dims
);
ctx
->
SetOutputDim
(
x_grad_name
,
x_dims
);
}
}
}
};
};
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录