Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
325ee637
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
325ee637
编写于
9月 26, 2017
作者:
Q
qiaolongfei
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix SoftmaxWithCrossEntropyOp
上级
dcfd31d7
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
38 addition
and
41 deletion
+38
-41
paddle/operators/math/softmax.cc
paddle/operators/math/softmax.cc
+1
-1
paddle/operators/softmax_with_cross_entropy_op.cc
paddle/operators/softmax_with_cross_entropy_op.cc
+37
-40
未找到文件。
paddle/operators/math/softmax.cc
浏览文件 @
325ee637
...
@@ -18,7 +18,7 @@ namespace paddle {
...
@@ -18,7 +18,7 @@ namespace paddle {
namespace
operators
{
namespace
operators
{
namespace
math
{
namespace
math
{
template
class
SoftmaxFunctor
<
platform
::
G
PUPlace
,
float
>;
template
class
SoftmaxFunctor
<
platform
::
C
PUPlace
,
float
>;
}
// namespace math
}
// namespace math
}
// namespace operators
}
// namespace operators
...
...
paddle/operators/softmax_with_cross_entropy_op.cc
浏览文件 @
325ee637
...
@@ -82,40 +82,38 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
...
@@ -82,40 +82,38 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
protected:
protected:
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContextBase
*
ctx
)
const
override
{
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"Logits"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Logits"
),
"Input(Logits) should be not null."
);
"Input(Logits) should be not null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"Label"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Label"
),
"Input(Label) should be not null."
);
"Input(Label) should be not null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Softmax"
),
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
OutputVar
(
"Softmax"
),
"Output(Softmax) should be not null."
);
"Output(Softmax) should be not null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Loss"
),
"Output(Loss) should be not null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
OutputVar
(
"Loss"
),
"Output(Loss) should be not null."
);
auto
logits_dims
=
ctx
->
GetInputDim
(
"Logits"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
const
Tensor
*
logits
=
ctx
.
Input
<
Tensor
>
(
"Logits"
);
const
Tensor
*
labels
=
ctx
.
Input
<
Tensor
>
(
"Label"
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
logits
->
dims
()
.
size
(),
2UL
,
logits
_dims
.
size
(),
2UL
,
"The input of softmax_with_cross_entropy should be a 2-D tensor."
);
"The input of softmax_with_cross_entropy should be a 2-D tensor."
);
PADDLE_ENFORCE_EQ
(
ctx
.
Input
<
Tensor
>
(
"Label"
)
->
dims
()
.
size
(),
2UL
,
PADDLE_ENFORCE_EQ
(
labels_dims
.
size
(),
2UL
,
"The labels should be a 2-D tensor."
);
"The labels should be a 2-D tensor."
);
if
(
ctx
.
Attr
<
bool
>
(
"softLabel"
))
{
if
(
ctx
->
Attrs
().
Get
<
bool
>
(
"softLabel"
))
{
PADDLE_ENFORCE_EQ
(
logits
->
dims
()[
1
],
labels
->
dims
()
[
1
],
PADDLE_ENFORCE_EQ
(
logits
_dims
[
1
],
labels_dims
[
1
],
"If Attr(softLabel) == true, the 2nd dimension of "
"If Attr(softLabel) == true, the 2nd dimension of "
"Input(X) and Input(Label) should be equal."
);
"Input(X) and Input(Label) should be equal."
);
}
else
{
}
else
{
PADDLE_ENFORCE_EQ
(
labels
->
dims
()
[
1
],
1UL
,
PADDLE_ENFORCE_EQ
(
labels
_dims
[
1
],
1UL
,
"If Attr(softLabel) == false, the 2nd dimension of "
"If Attr(softLabel) == false, the 2nd dimension of "
"Input(Label) should be 1."
);
"Input(Label) should be 1."
);
}
}
ctx
.
Output
<
framework
::
Tensor
>
(
"Softmax"
)
->
Resize
(
logits
->
dims
()
);
ctx
->
SetOutputDim
(
"Softmax"
,
logits_dims
);
ctx
.
Output
<
framework
::
Tensor
>
(
"Loss"
)
->
Resize
({
logits
->
dims
()
[
0
],
1
});
ctx
->
SetOutputDim
(
"Loss"
,
{
logits_dims
[
0
],
1
});
ctx
.
ShareLoD
(
"Logits"
,
/*->*/
"Softmax"
);
ctx
->
ShareLoD
(
"Logits"
,
/*->*/
"Softmax"
);
ctx
.
ShareLoD
(
"Logits"
,
/*->*/
"Loss"
);
ctx
->
ShareLoD
(
"Logits"
,
/*->*/
"Loss"
);
}
}
};
};
...
@@ -124,33 +122,32 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
...
@@ -124,33 +122,32 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
protected:
protected:
void
InferShape
(
const
framework
::
InferShapeContext
&
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContextBase
*
ctx
)
const
override
{
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
framework
::
GradVarName
(
"Loss"
)),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Loss"
)),
"Input(Loss@Grad) should not be null."
);
"Input(Loss@Grad) should not be null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"Softmax"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Softmax"
),
"Input(Softmax) should be not null."
);
"Input(Softmax) should be not null."
);
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
InputVar
(
"Label"
),
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Label"
),
"Input(Label) should be not null."
);
"Input(Label) should be not null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"Logits"
)),
PADDLE_ENFORCE_NOT_NULL
(
ctx
.
OutputVar
(
framework
::
GradVarName
(
"Logits"
)),
"Output(Logits@Grad) should be not null."
);
"Output(Logits@Grad) should be not null."
);
auto
softmax_dims
=
ctx
->
GetInputDim
(
"Softmax"
);
const
Tensor
*
softmax
=
ctx
.
Input
<
Tensor
>
(
"Softmax"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
const
Tensor
*
labels
=
ctx
.
Input
<
Tensor
>
(
"Label"
);
PADDLE_ENFORCE_EQ
(
labels_dims
.
size
(),
2UL
,
PADDLE_ENFORCE_EQ
(
ctx
.
Input
<
Tensor
>
(
"Label"
)
->
dims
().
size
(),
2UL
,
"The labels should be a 2-D tensor."
);
"The labels should be a 2-D tensor."
);
if
(
ctx
.
Attr
<
bool
>
(
"softLabel"
))
{
if
(
ctx
->
Attrs
().
Get
<
bool
>
(
"softLabel"
))
{
PADDLE_ENFORCE_EQ
(
softmax
->
dims
()[
1
],
labels
->
dims
()
[
1
],
PADDLE_ENFORCE_EQ
(
softmax
_dims
[
1
],
labels_dims
[
1
],
"When Attr(softLabel) == true, the 2nd dimension of "
"When Attr(softLabel) == true, the 2nd dimension of "
"Input(X) and Input(Label) should be equal."
);
"Input(X) and Input(Label) should be equal."
);
}
else
{
}
else
{
PADDLE_ENFORCE_EQ
(
labels
->
dims
()
[
1
],
1UL
,
PADDLE_ENFORCE_EQ
(
labels
_dims
[
1
],
1UL
,
"When Attr(softLabel) == false, the 2nd dimension of "
"When Attr(softLabel) == false, the 2nd dimension of "
"Input(Label) should be 1."
);
"Input(Label) should be 1."
);
}
}
ctx
.
Output
<
framework
::
LoDTensor
>
(
framework
::
GradVarName
(
"Logits"
))
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"Logits"
),
->
Resize
(
ctx
.
Input
<
Tensor
>
(
"Softmax"
)
->
dims
(
));
ctx
->
GetInputDim
(
"Softmax"
));
}
}
};
};
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录