Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
28d42a94
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
28d42a94
编写于
5月 06, 2021
作者:
C
chajchaj
提交者:
chajchaj
5月 06, 2021
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
change parameter name from softmax_switch to use_softmax, test=develop
上级
2fe45806
变更
5
隐藏空白更改
内联
并排
Showing
5 changed file
with
56 addition
and
59 deletion
+56
-59
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
+3
-4
paddle/fluid/operators/softmax_with_cross_entropy_op.cu
paddle/fluid/operators/softmax_with_cross_entropy_op.cu
+4
-4
paddle/fluid/operators/softmax_with_cross_entropy_op.h
paddle/fluid/operators/softmax_with_cross_entropy_op.h
+8
-8
python/paddle/fluid/tests/unittests/test_softmax_with_cross_entropy_op.py
...uid/tests/unittests/test_softmax_with_cross_entropy_op.py
+39
-39
python/paddle/nn/functional/loss.py
python/paddle/nn/functional/loss.py
+2
-4
未找到文件。
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
浏览文件 @
28d42a94
...
...
@@ -55,7 +55,7 @@ class SoftmaxWithCrossEntropyOpMaker
"the given labels as soft labels."
)
.
SetDefault
(
false
);
AddAttr
<
bool
>
(
"
softmax_switch
"
,
"
use_softmax
"
,
"(bool, default: true), A flag to indicate whether to do softmax "
)
.
SetDefault
(
true
);
AddAttr
<
bool
>
(
...
...
@@ -320,7 +320,6 @@ REGISTER_OP_CPU_KERNEL(softmax_with_cross_entropy_grad,
REGISTER_OP_VERSION
(
softmax_with_cross_entropy
)
.
AddCheckpoint
(
R"ROC(
Add a new attribute [
softmax_switch
] )ROC"
,
Add a new attribute [
use_softmax
] )ROC"
,
paddle
::
framework
::
compatible
::
OpVersionDesc
().
NewAttr
(
"softmax_switch"
,
"A flag to indicate whether to do softmax"
,
true
));
"use_softmax"
,
"A flag to indicate whether to do softmax"
,
true
));
paddle/fluid/operators/softmax_with_cross_entropy_op.cu
浏览文件 @
28d42a94
...
...
@@ -772,10 +772,10 @@ class SoftmaxWithCrossEntropyCUDAKernel : public framework::OpKernel<T> {
platform
::
is_gpu_place
(
context
.
GetPlace
()),
true
,
platform
::
errors
::
Unavailable
(
"softmax_with_cross_entropy operator's "
"CUDA kernel only runs on GPU device."
));
const
bool
softmax_switch
=
context
.
Attr
<
bool
>
(
"softmax_switch
"
);
const
bool
use_softmax
=
context
.
Attr
<
bool
>
(
"use_softmax
"
);
// do not with softmax op, and input is softmax
if
(
!
softmax_switch
)
{
if
(
!
use_softmax
)
{
const
Tensor
*
softmax
=
context
.
Input
<
Tensor
>
(
"Logits"
);
const
Tensor
*
labels
=
context
.
Input
<
Tensor
>
(
"Label"
);
Tensor
*
softmax_out
=
context
.
Output
<
Tensor
>
(
"Softmax"
);
...
...
@@ -925,10 +925,10 @@ class SoftmaxWithCrossEntropyGradCUDAKernel : public framework::OpKernel<T> {
int
block
=
512
;
auto
stream
=
context
.
cuda_device_context
().
stream
();
auto
ignore_index
=
context
.
Attr
<
int
>
(
"ignore_index"
);
auto
softmax_switch
=
context
.
Attr
<
bool
>
(
"softmax_switch
"
);
auto
use_softmax
=
context
.
Attr
<
bool
>
(
"use_softmax
"
);
// do not with softmax op, and input is softmax
if
(
!
softmax_switch
)
{
if
(
!
use_softmax
)
{
if
(
context
.
Attr
<
bool
>
(
"soft_label"
))
{
int
grid
=
(
n
*
d
+
block
-
1
)
/
block
;
const
T
*
label_data
=
labels
->
data
<
T
>
();
...
...
paddle/fluid/operators/softmax_with_cross_entropy_op.h
浏览文件 @
28d42a94
...
...
@@ -31,10 +31,10 @@ class SoftmaxWithCrossEntropyKernel : public framework::OpKernel<T> {
PADDLE_ENFORCE_EQ
(
platform
::
is_cpu_place
(
context
.
GetPlace
()),
true
,
platform
::
errors
::
Unimplemented
(
"This kernel only runs on CPU."
));
const
bool
softmax_switch
=
context
.
Attr
<
bool
>
(
"softmax_switch
"
);
const
bool
use_softmax
=
context
.
Attr
<
bool
>
(
"use_softmax
"
);
// do not with softmax op, and input is softmax
if
(
!
softmax_switch
)
{
if
(
!
use_softmax
)
{
const
Tensor
*
softmax
=
context
.
Input
<
Tensor
>
(
"Logits"
);
const
Tensor
*
labels
=
context
.
Input
<
Tensor
>
(
"Label"
);
Tensor
*
softmax_out
=
context
.
Output
<
Tensor
>
(
"Softmax"
);
...
...
@@ -113,9 +113,9 @@ class SoftmaxWithCrossEntropyGradKernel : public framework::OpKernel<T> {
context
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"Logits"
));
const
Tensor
*
softmax
=
context
.
Input
<
Tensor
>
(
"Softmax"
);
const
bool
softmax_switch
=
context
.
Attr
<
bool
>
(
"softmax_switch
"
);
const
bool
use_softmax
=
context
.
Attr
<
bool
>
(
"use_softmax
"
);
if
(
logit_grad
!=
softmax
||
!
softmax_switch
)
{
if
(
logit_grad
!=
softmax
||
!
use_softmax
)
{
framework
::
TensorCopy
(
*
softmax
,
context
.
GetPlace
(),
context
.
device_context
(),
logit_grad
);
}
...
...
@@ -138,8 +138,8 @@ class SoftmaxWithCrossEntropyGradKernel : public framework::OpKernel<T> {
auto
logit_grad_mat
=
framework
::
EigenMatrix
<
T
>::
From
(
logit_grad_2d
);
auto
&
place
=
*
context
.
template
device_context
<
platform
::
CPUDeviceContext
>()
.
eigen_device
();
if
(
!
softmax_switch
)
{
//
softmax_switch
step1
if
(
!
use_softmax
)
{
//
use_softmax
step1
if
(
soft_label
)
{
auto
lbl_mat
=
framework
::
EigenMatrix
<
T
>::
From
(
labels_2d
);
logit_grad_mat
.
device
(
place
)
=
...
...
@@ -148,7 +148,7 @@ class SoftmaxWithCrossEntropyGradKernel : public framework::OpKernel<T> {
out_grad_mat
.
broadcast
(
Eigen
::
DSizes
<
int
,
2
>
(
1
,
axis_dim
))
*
logit_grad_mat
;
}
//
softmax_switch
step2
//
use_softmax
step2
else
{
const
int64_t
*
label_data
=
labels
->
data
<
int64_t
>
();
T
*
logit_grad_data
=
logit_grad
->
data
<
T
>
();
...
...
@@ -181,7 +181,7 @@ class SoftmaxWithCrossEntropyGradKernel : public framework::OpKernel<T> {
return
;
}
// for
softmax_switch
=False, continue
// for
use_softmax
=False, continue
if
(
soft_label
)
{
// when soft_label = True, ignore_index is not supported
...
...
python/paddle/fluid/tests/unittests/test_softmax_with_cross_entropy_op.py
浏览文件 @
28d42a94
...
...
@@ -56,7 +56,7 @@ class TestSoftmaxWithCrossEntropyOp(OpTest):
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
41
,
37
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
def
setUp
(
self
):
self
.
initParams
()
...
...
@@ -77,7 +77,7 @@ class TestSoftmaxWithCrossEntropyOp(OpTest):
loss
=
cross_entropy
(
softmax
,
labels
,
self
.
soft_label
,
self
.
axis
,
self
.
ignore_index
)
if
self
.
softmax_switch
==
False
:
if
self
.
use_softmax
==
False
:
self
.
inputs
=
{
"Logits"
:
softmax
,
"Label"
:
labels
}
else
:
self
.
inputs
=
{
"Logits"
:
logits
,
"Label"
:
labels
}
...
...
@@ -90,7 +90,7 @@ class TestSoftmaxWithCrossEntropyOp(OpTest):
"numeric_stable_mode"
:
self
.
numeric_stable_mode
,
"soft_label"
:
self
.
soft_label
,
"ignore_index"
:
self
.
ignore_index
,
"
softmax_switch"
:
self
.
softmax_switch
,
"
use_softmax"
:
self
.
use_softmax
,
}
if
self
.
axis
!=
-
1
:
...
...
@@ -117,7 +117,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_1D(
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_1D
(
...
...
@@ -130,7 +130,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_1D(
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
##############################################################################
...
...
@@ -146,7 +146,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D(
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis2
(
...
...
@@ -159,7 +159,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis2(
self
.
axis
=
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis3
(
...
...
@@ -172,7 +172,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis3(
self
.
axis
=
2
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis4
(
...
...
@@ -185,7 +185,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_SoftLabel_2D_Axis4(
self
.
axis
=
3
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
##############################################################################
...
...
@@ -207,7 +207,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D(
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis2
(
...
...
@@ -220,7 +220,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis2(
self
.
axis
=
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis3
(
...
...
@@ -233,7 +233,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis3(
self
.
axis
=
2
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis4
(
...
...
@@ -246,7 +246,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Axis4(
self
.
axis
=
3
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
##############################################################################
...
...
@@ -268,7 +268,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_Ignore(
self
.
axis
=
-
1
self
.
ignore_index
=
2
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_Ignore_Axis
(
...
...
@@ -281,7 +281,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_Ignore_Axis(
self
.
axis
=
1
self
.
ignore_index
=
2
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Ignore
(
...
...
@@ -294,7 +294,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Ignore(
self
.
axis
=
-
1
self
.
ignore_index
=
2
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
class
TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Ignore_Axis3
(
...
...
@@ -307,7 +307,7 @@ class TestSoftmaxWithCrossEntropyOp_NotWithSoftmax_HardLabel_2D_Ignore_Axis3(
self
.
axis
=
2
self
.
ignore_index
=
2
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
False
#default is true, means "with softmax"
self
.
use_softmax
=
False
#default is true, means "with softmax"
##############################################################################
...
...
@@ -324,7 +324,7 @@ class TestSoftmaxWithCrossEntropyOpNoCudnn(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
@
unittest
.
skipIf
(
not
core
.
is_compiled_with_cuda
(),
...
...
@@ -403,7 +403,7 @@ class TestSoftmaxWithCrossEntropyOp2(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
41
,
37
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
def
test_check_output
(
self
):
self
.
check_output
()
...
...
@@ -429,7 +429,7 @@ class TestSoftmaxWithCrossEntropyOp3(TestSoftmaxWithCrossEntropyOp):
self
.
ignore_index
=
5
self
.
axis
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOp3NoCudnn
(
TestSoftmaxWithCrossEntropyOp3
):
...
...
@@ -441,7 +441,7 @@ class TestSoftmaxWithCrossEntropyOp3NoCudnn(TestSoftmaxWithCrossEntropyOp3):
self
.
ignore_index
=
4
self
.
axis
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpAxis1
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -458,7 +458,7 @@ class TestSoftmaxWithCrossEntropyOpAxis1(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
0
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpAxis2
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -475,7 +475,7 @@ class TestSoftmaxWithCrossEntropyOpAxis2(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpAxis3
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -492,7 +492,7 @@ class TestSoftmaxWithCrossEntropyOpAxis3(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
2
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpAxis4
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -509,7 +509,7 @@ class TestSoftmaxWithCrossEntropyOpAxis4(TestSoftmaxWithCrossEntropyOp):
self
.
axis
=
3
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
11
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpAxisDimEqualOne
(
...
...
@@ -527,7 +527,7 @@ class TestSoftmaxWithCrossEntropyOpAxisDimEqualOne(
self
.
axis
=
-
1
self
.
ignore_index
=
-
1
self
.
shape
=
[
3
,
5
,
7
,
1
]
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis1
(
...
...
@@ -540,7 +540,7 @@ class TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis1(
self
.
axis
=
0
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float16
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis2
(
...
...
@@ -553,7 +553,7 @@ class TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis2(
self
.
axis
=
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float16
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis3
(
...
...
@@ -566,7 +566,7 @@ class TestSoftmaxWithCrossEntropyOpNoCudnnFp16Axis3(
self
.
axis
=
2
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float16
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpSoftLabelAxis1
(
...
...
@@ -579,7 +579,7 @@ class TestSoftmaxWithCrossEntropyOpSoftLabelAxis1(
self
.
axis
=
0
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpSoftLabelAxis2
(
...
...
@@ -592,7 +592,7 @@ class TestSoftmaxWithCrossEntropyOpSoftLabelAxis2(
self
.
axis
=
1
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpSoftLabelAxis3
(
...
...
@@ -605,7 +605,7 @@ class TestSoftmaxWithCrossEntropyOpSoftLabelAxis3(
self
.
axis
=
2
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpSoftLabelAxis4
(
...
...
@@ -618,7 +618,7 @@ class TestSoftmaxWithCrossEntropyOpSoftLabelAxis4(
self
.
axis
=
3
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis1
(
...
...
@@ -631,7 +631,7 @@ class TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis1(
self
.
ignore_index
=
1
self
.
axis
=
0
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis2
(
...
...
@@ -644,7 +644,7 @@ class TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis2(
self
.
ignore_index
=
0
self
.
axis
=
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis3
(
...
...
@@ -657,7 +657,7 @@ class TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis3(
self
.
ignore_index
=
3
self
.
axis
=
2
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis4
(
...
...
@@ -670,7 +670,7 @@ class TestSoftmaxWithCrossEntropyOpIgnoreIndexNoCudnnAxis4(
self
.
ignore_index
=
3
self
.
axis
=
3
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpBoundary0
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -688,7 +688,7 @@ class TestSoftmaxWithCrossEntropyOpBoundary0(TestSoftmaxWithCrossEntropyOp):
self
.
ignore_index
=
-
1
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
logits
=
np
.
full
(
self
.
shape
,
-
500.0
).
astype
(
self
.
dtype
)
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
class
TestSoftmaxWithCrossEntropyOpBoundary1
(
TestSoftmaxWithCrossEntropyOp
):
...
...
@@ -707,7 +707,7 @@ class TestSoftmaxWithCrossEntropyOpBoundary1(TestSoftmaxWithCrossEntropyOp):
self
.
dtype
=
np
.
float32
if
core
.
is_compiled_with_rocm
()
else
np
.
float64
self
.
logits
=
np
.
full
(
self
.
shape
,
1000.0
).
astype
(
self
.
dtype
)
self
.
logits
[:,
:,
0
,
:]
=
-
1000.0
self
.
softmax_switch
=
True
self
.
use_softmax
=
True
if
__name__
==
"__main__"
:
...
...
python/paddle/nn/functional/loss.py
浏览文件 @
28d42a94
...
...
@@ -1371,8 +1371,6 @@ def cross_entropy(input,
"should be '-100', but received %s, which is not allowed."
%
ignore_index
)
softmax_switch
=
use_softmax
input_dims
=
len
(
list
(
input
.
shape
))
label_dims
=
len
(
list
(
label
.
shape
))
if
input_dims
-
1
!=
label_dims
and
input_dims
!=
label_dims
:
...
...
@@ -1385,7 +1383,7 @@ def cross_entropy(input,
_
,
out
=
core
.
ops
.
softmax_with_cross_entropy
(
input
,
label
,
'soft_label'
,
soft_label
,
'ignore_index'
,
ignore_index
,
'numeric_stable_mode'
,
True
,
'axis'
,
axis
,
'
softmax_switch'
,
softmax_switch
)
'
use_softmax'
,
use_softmax
)
if
weight
is
not
None
:
...
...
@@ -1467,7 +1465,7 @@ def cross_entropy(input,
'ignore_index'
:
ignore_index
,
'numeric_stable_mode'
:
True
,
'axis'
:
axis
,
'
softmax_switch'
:
softmax_switch
'
use_softmax'
:
use_softmax
}
helper
=
LayerHelper
(
'softmax_with_cross_entropy'
,
**
locals
())
softmax
=
helper
.
create_variable_for_type_inference
(
dtype
=
input
.
dtype
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录