Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
05c9642d
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
05c9642d
编写于
5月 14, 2020
作者:
S
suytingwan
提交者:
GitHub
5月 14, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Update paddle enforce message (#24498)
* test=develop error message update
上级
9f83f0fe
变更
19
隐藏空白更改
内联
并排
Showing
19 changed file
with
534 addition
and
281 deletion
+534
-281
paddle/fluid/operators/bilinear_tensor_product_op.cc
paddle/fluid/operators/bilinear_tensor_product_op.cc
+62
-31
paddle/fluid/operators/detection/anchor_generator_op.cc
paddle/fluid/operators/detection/anchor_generator_op.cc
+27
-14
paddle/fluid/operators/detection/bipartite_match_op.cc
paddle/fluid/operators/detection/bipartite_match_op.cc
+30
-14
paddle/fluid/operators/detection/generate_mask_labels_op.cc
paddle/fluid/operators/detection/generate_mask_labels_op.cc
+73
-35
paddle/fluid/operators/detection/target_assign_op.cc
paddle/fluid/operators/detection/target_assign_op.cc
+26
-14
paddle/fluid/operators/detection/target_assign_op.h
paddle/fluid/operators/detection/target_assign_op.h
+7
-2
paddle/fluid/operators/filter_by_instag_op.cc
paddle/fluid/operators/filter_by_instag_op.cc
+25
-14
paddle/fluid/operators/one_hot_v2_op.h
paddle/fluid/operators/one_hot_v2_op.h
+11
-3
paddle/fluid/operators/pool_with_index_op.cc
paddle/fluid/operators/pool_with_index_op.cc
+27
-19
paddle/fluid/operators/psroi_pool_op.cc
paddle/fluid/operators/psroi_pool_op.cc
+43
-27
paddle/fluid/operators/psroi_pool_op.h
paddle/fluid/operators/psroi_pool_op.h
+8
-4
paddle/fluid/operators/roi_pool_op.cc
paddle/fluid/operators/roi_pool_op.cc
+4
-1
paddle/fluid/operators/roi_pool_op.h
paddle/fluid/operators/roi_pool_op.h
+8
-4
paddle/fluid/operators/softmax_op.cc
paddle/fluid/operators/softmax_op.cc
+34
-18
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
+70
-39
paddle/fluid/operators/softmax_with_cross_entropy_op.h
paddle/fluid/operators/softmax_with_cross_entropy_op.h
+3
-2
paddle/fluid/operators/spp_op.cc
paddle/fluid/operators/spp_op.cc
+15
-10
paddle/fluid/operators/unsqueeze_op.cc
paddle/fluid/operators/unsqueeze_op.cc
+51
-27
paddle/fluid/operators/unsqueeze_op.h
paddle/fluid/operators/unsqueeze_op.h
+10
-3
未找到文件。
paddle/fluid/operators/bilinear_tensor_product_op.cc
浏览文件 @
05c9642d
...
@@ -28,39 +28,61 @@ class BilinearTensorProductOp : public framework::OperatorWithKernel {
...
@@ -28,39 +28,61 @@ class BilinearTensorProductOp : public framework::OperatorWithKernel {
protected:
protected:
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
"Input(X) should not be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Y"
),
"Input(Y) should not be null."
);
ctx
->
HasInput
(
"X"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Weight"
),
platform
::
errors
::
InvalidArgument
(
"Input(X) should not be null."
));
"Input(Weight) should not be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
"Output(Out) should not be null."
);
ctx
->
HasInput
(
"Y"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Y) should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Weight"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Weight) should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Out) should not be null."
));
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
y_dims
=
ctx
->
GetInputDim
(
"Y"
);
auto
y_dims
=
ctx
->
GetInputDim
(
"Y"
);
auto
weight_dims
=
ctx
->
GetInputDim
(
"Weight"
);
auto
weight_dims
=
ctx
->
GetInputDim
(
"Weight"
);
PADDLE_ENFORCE_EQ
(
x_dims
.
size
(),
2UL
,
"The input(X) must be a 2D Tensor."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
y_dims
.
size
(),
2UL
,
"The input(Y) must be a 2D Tensor."
);
x_dims
.
size
(),
2UL
,
platform
::
errors
::
InvalidArgument
(
"The input(X) must be a 2D Tensor."
));
PADDLE_ENFORCE_EQ
(
y_dims
.
size
(),
2UL
,
platform
::
errors
::
InvalidArgument
(
"The input(Y) must be a 2D Tensor."
));
PADDLE_ENFORCE_EQ
(
weight_dims
.
size
(),
3UL
,
PADDLE_ENFORCE_EQ
(
weight_dims
.
size
(),
3UL
,
"The input(Weight) must be a 3D tensor."
);
platform
::
errors
::
InvalidArgument
(
"The input(Weight) must be a 3D tensor."
));
if
(
ctx
->
IsRuntime
()
||
(
x_dims
[
0
]
>
0
&&
y_dims
[
0
]
>
0
))
{
if
(
ctx
->
IsRuntime
()
||
(
x_dims
[
0
]
>
0
&&
y_dims
[
0
]
>
0
))
{
PADDLE_ENFORCE_EQ
(
x_dims
[
0
],
y_dims
[
0
],
PADDLE_ENFORCE_EQ
(
"The first dimension(batch_size) of input(X) must be "
x_dims
[
0
],
y_dims
[
0
],
"equal to the first dimension of the input(Y)."
);
platform
::
errors
::
InvalidArgument
(
"The first dimension(batch_size) of input(X) must be "
"equal to the first dimension of the input(Y)."
));
}
}
PADDLE_ENFORCE_EQ
(
x_dims
[
1
],
weight_dims
[
1
],
PADDLE_ENFORCE_EQ
(
x_dims
[
1
],
weight_dims
[
1
],
"The second dimension of input(X) must be equal to "
platform
::
errors
::
InvalidArgument
(
"the second dimension of the input(Weight)."
);
"The second dimension of input(X) must be equal to "
"the second dimension of the input(Weight)."
));
PADDLE_ENFORCE_EQ
(
y_dims
[
1
],
weight_dims
[
2
],
PADDLE_ENFORCE_EQ
(
y_dims
[
1
],
weight_dims
[
2
],
"The second dimension of input(Y) must be equal to "
platform
::
errors
::
InvalidArgument
(
"the third dimension of the input(Weight)."
);
"The second dimension of input(Y) must be equal to "
"the third dimension of the input(Weight)."
));
if
(
ctx
->
HasInput
(
"Bias"
))
{
if
(
ctx
->
HasInput
(
"Bias"
))
{
auto
bias_dims
=
ctx
->
GetInputDim
(
"Bias"
);
auto
bias_dims
=
ctx
->
GetInputDim
(
"Bias"
);
PADDLE_ENFORCE
(
bias_dims
.
size
()
==
2UL
&&
bias_dims
[
0
]
==
1UL
,
PADDLE_ENFORCE_EQ
(
bias_dims
.
size
(),
2UL
,
"The Input(Bias) must be a 2-D tensor with "
platform
::
errors
::
InvalidArgument
(
"the 2nd dimension fixed to 1 (a row vector)."
);
"The Input(Bias) must be a 2-D tensor with "
"the 2nd dimension fixed to 1 (a row vector)."
));
PADDLE_ENFORCE_EQ
(
bias_dims
[
0
],
1UL
,
platform
::
errors
::
InvalidArgument
(
"The Input(Bias) must be a 2-D tensor with "
"the 2nd dimension fixed to 1 (a row vector)."
));
PADDLE_ENFORCE_EQ
(
bias_dims
[
1
],
weight_dims
[
0
],
PADDLE_ENFORCE_EQ
(
bias_dims
[
1
],
weight_dims
[
0
],
"The second dimension of input(Bias) must be equal "
platform
::
errors
::
InvalidArgument
(
"to the first dimension of the input(Weight)."
);
"The second dimension of input(Bias) must be equal "
"to the first dimension of the input(Weight)."
));
}
}
ctx
->
SetOutputDim
(
"Out"
,
{
x_dims
[
0
],
weight_dims
[
0
]});
ctx
->
SetOutputDim
(
"Out"
,
{
x_dims
[
0
],
weight_dims
[
0
]});
...
@@ -104,27 +126,36 @@ class BilinearTensorProductOpGrad : public framework::OperatorWithKernel {
...
@@ -104,27 +126,36 @@ class BilinearTensorProductOpGrad : public framework::OperatorWithKernel {
protected:
protected:
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
"Input(X) should not be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Y"
),
"Input(Y) should not be null."
);
ctx
->
HasInput
(
"X"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Weight"
),
platform
::
errors
::
InvalidArgument
(
"Input(X) should not be null."
));
"Input(Weight) should not be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
ctx
->
HasInput
(
"Y"
),
true
,
"Input(Out@GRAD) should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Y) should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Weight"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Weight) should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Out@GRAD) should not be null."
));
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
y_dims
=
ctx
->
GetInputDim
(
"Y"
);
auto
y_dims
=
ctx
->
GetInputDim
(
"Y"
);
auto
weight_dims
=
ctx
->
GetInputDim
(
"Weight"
);
auto
weight_dims
=
ctx
->
GetInputDim
(
"Weight"
);
auto
out_dims
=
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
));
auto
out_dims
=
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
));
PADDLE_ENFORCE_EQ
(
out_dims
.
size
(),
2UL
,
PADDLE_ENFORCE_EQ
(
out_dims
.
size
(),
2UL
,
"The input(Out@GRAD) must be a 2D Tensor."
);
platform
::
errors
::
InvalidArgument
(
"The input(Out@GRAD) must be a 2D Tensor."
));
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
x_dims
[
0
],
out_dims
[
0
],
x_dims
[
0
],
out_dims
[
0
],
"The first dimension(batch_size) of input(Out@GRAD) must be "
platform
::
errors
::
InvalidArgument
(
"equal to the first dimension of the Input(X)."
);
"The first dimension(batch_size) of input(Out@GRAD) must be "
"equal to the first dimension of the Input(X)."
));
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
weight_dims
[
0
],
out_dims
[
1
],
weight_dims
[
0
],
out_dims
[
1
],
"The second dimension of input(Out@GRAD) must be equal to "
platform
::
errors
::
InvalidArgument
(
"the third dimension of the Input(Weight)."
);
"The second dimension of input(Out@GRAD) must be equal to "
"the third dimension of the Input(Weight)."
));
auto
bias_grad_name
=
framework
::
GradVarName
(
"Bias"
);
auto
bias_grad_name
=
framework
::
GradVarName
(
"Bias"
);
if
(
ctx
->
HasOutput
(
bias_grad_name
))
{
if
(
ctx
->
HasOutput
(
bias_grad_name
))
{
...
...
paddle/fluid/operators/detection/anchor_generator_op.cc
浏览文件 @
05c9642d
...
@@ -22,16 +22,23 @@ class AnchorGeneratorOp : public framework::OperatorWithKernel {
...
@@ -22,16 +22,23 @@ class AnchorGeneratorOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Input"
),
PADDLE_ENFORCE_EQ
(
"Input(Input) of AnchorGeneratorOp should not be null."
);
ctx
->
HasInput
(
"Input"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Anchors"
),
platform
::
errors
::
InvalidArgument
(
"Output(Anchors) of AnchorGeneratorOp should not be null."
);
"Input(Input) of AnchorGeneratorOp should not be null."
));
PADDLE_ENFORCE
(
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Variances"
),
ctx
->
HasOutput
(
"Anchors"
),
true
,
"Output(Variances) of AnchorGeneratorOp should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Output(Anchors) of AnchorGeneratorOp should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Variances"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Variances) of AnchorGeneratorOp should not be null."
));
auto
input_dims
=
ctx
->
GetInputDim
(
"Input"
);
auto
input_dims
=
ctx
->
GetInputDim
(
"Input"
);
PADDLE_ENFORCE
(
input_dims
.
size
()
==
4
,
"The layout of input is NCHW."
);
PADDLE_ENFORCE_EQ
(
input_dims
.
size
(),
4
,
platform
::
errors
::
InvalidArgument
(
"The layout of input is NCHW."
));
auto
anchor_sizes
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
float
>>
(
"anchor_sizes"
);
auto
anchor_sizes
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
float
>>
(
"anchor_sizes"
);
auto
aspect_ratios
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
float
>>
(
"aspect_ratios"
);
auto
aspect_ratios
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
float
>>
(
"aspect_ratios"
);
...
@@ -87,10 +94,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -87,10 +94,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
"equals to 64**2."
)
"equals to 64**2."
)
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
anchor_sizes
)
{
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
anchor_sizes
)
{
PADDLE_ENFORCE_GT
(
anchor_sizes
.
size
(),
0UL
,
PADDLE_ENFORCE_GT
(
anchor_sizes
.
size
(),
0UL
,
"Size of anchor_sizes must be at least 1."
);
platform
::
errors
::
InvalidArgument
(
"Size of anchor_sizes must be at least 1."
));
for
(
size_t
i
=
0
;
i
<
anchor_sizes
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
anchor_sizes
.
size
();
++
i
)
{
PADDLE_ENFORCE_GT
(
anchor_sizes
[
i
],
0.0
,
PADDLE_ENFORCE_GT
(
anchor_sizes
[
i
],
0.0
,
"anchor_sizes[%d] must be positive."
,
i
);
platform
::
errors
::
InvalidArgument
(
"anchor_sizes[%d] must be positive."
,
i
));
}
}
});
});
AddAttr
<
std
::
vector
<
float
>>
(
AddAttr
<
std
::
vector
<
float
>>
(
...
@@ -105,10 +114,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -105,10 +114,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
"in box regression deltas"
)
"in box regression deltas"
)
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
variances
)
{
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
variances
)
{
PADDLE_ENFORCE_EQ
(
variances
.
size
(),
4UL
,
PADDLE_ENFORCE_EQ
(
variances
.
size
(),
4UL
,
"Must and only provide 4 variance."
);
platform
::
errors
::
InvalidArgument
(
"Must provide 4 variance only."
));
for
(
size_t
i
=
0
;
i
<
variances
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
variances
.
size
();
++
i
)
{
PADDLE_ENFORCE_GT
(
variances
[
i
],
0.0
,
PADDLE_ENFORCE_GT
(
variances
[
i
],
0.0
,
"variance[%d] must be greater than 0."
,
i
);
platform
::
errors
::
InvalidArgument
(
"variance[%d] must be greater than 0."
,
i
));
}
}
});
});
...
@@ -119,10 +130,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -119,10 +130,12 @@ class AnchorGeneratorOpMaker : public framework::OpProtoAndCheckerMaker {
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
stride
)
{
.
AddCustomChecker
([](
const
std
::
vector
<
float
>&
stride
)
{
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
stride
.
size
(),
2UL
,
stride
.
size
(),
2UL
,
"Must and only provide 2 stride for width and height."
);
platform
::
errors
::
InvalidArgument
(
"Must provide 2 stride for width and height only."
));
for
(
size_t
i
=
0
;
i
<
stride
.
size
();
++
i
)
{
for
(
size_t
i
=
0
;
i
<
stride
.
size
();
++
i
)
{
PADDLE_ENFORCE_GT
(
stride
[
i
],
0.0
,
PADDLE_ENFORCE_GT
(
stride
[
i
],
0.0
,
"stride[%d] should be larger than 0."
,
i
);
platform
::
errors
::
InvalidArgument
(
"stride[%d] should be larger than 0."
,
i
));
}
}
});
});
...
...
paddle/fluid/operators/detection/bipartite_match_op.cc
浏览文件 @
05c9642d
...
@@ -26,17 +26,23 @@ class BipartiteMatchOp : public framework::OperatorWithKernel {
...
@@ -26,17 +26,23 @@ class BipartiteMatchOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"DistMat"
),
PADDLE_ENFORCE_EQ
(
"Input(DistMat) of BipartiteMatch should not be null."
);
ctx
->
HasInput
(
"DistMat"
),
true
,
PADDLE_ENFORCE
(
platform
::
errors
::
InvalidArgument
(
ctx
->
HasOutput
(
"ColToRowMatchIndices"
),
"Input(DistMat) of BipartiteMatch should not be null."
));
"Output(ColToRowMatchIndices) of BipartiteMatch should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"ColToRowMatchIndices"
),
true
,
PADDLE_ENFORCE
(
platform
::
errors
::
InvalidArgument
(
ctx
->
HasOutput
(
"ColToRowMatchDist"
),
"Output(ColToRowMatchIndices) of BipartiteMatch "
"Output(ColToRowMatchDist) of BipartiteMatch should not be null."
);
"should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"ColToRowMatchDist"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(ColToRowMatchDist) of BipartiteMatch should not be null."
));
auto
dims
=
ctx
->
GetInputDim
(
"DistMat"
);
auto
dims
=
ctx
->
GetInputDim
(
"DistMat"
);
PADDLE_ENFORCE_EQ
(
dims
.
size
(),
2
,
"The rank of Input(DistMat) must be 2."
);
PADDLE_ENFORCE_EQ
(
dims
.
size
(),
2
,
platform
::
errors
::
InvalidArgument
(
"The rank of Input(DistMat) must be 2."
));
ctx
->
SetOutputDim
(
"ColToRowMatchIndices"
,
dims
);
ctx
->
SetOutputDim
(
"ColToRowMatchIndices"
,
dims
);
ctx
->
SetOutputDim
(
"ColToRowMatchDist"
,
dims
);
ctx
->
SetOutputDim
(
"ColToRowMatchDist"
,
dims
);
...
@@ -64,7 +70,9 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
...
@@ -64,7 +70,9 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
// The match_dist must be initialized to 0 at first.
// The match_dist must be initialized to 0 at first.
void
BipartiteMatch
(
const
Tensor
&
dist
,
int
*
match_indices
,
void
BipartiteMatch
(
const
Tensor
&
dist
,
int
*
match_indices
,
T
*
match_dist
)
const
{
T
*
match_dist
)
const
{
PADDLE_ENFORCE_EQ
(
dist
.
dims
().
size
(),
2
,
"The rank of dist must be 2."
);
PADDLE_ENFORCE_EQ
(
dist
.
dims
().
size
(),
2
,
platform
::
errors
::
InvalidArgument
(
"The rank of dist must be 2."
));
int64_t
row
=
dist
.
dims
()[
0
];
int64_t
row
=
dist
.
dims
()[
0
];
int64_t
col
=
dist
.
dims
()[
1
];
int64_t
col
=
dist
.
dims
()[
1
];
auto
*
dist_data
=
dist
.
data
<
T
>
();
auto
*
dist_data
=
dist
.
data
<
T
>
();
...
@@ -127,7 +135,11 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
...
@@ -127,7 +135,11 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
// Cannot find good match.
// Cannot find good match.
break
;
break
;
}
else
{
}
else
{
PADDLE_ENFORCE_EQ
(
match_indices
[
max_idx
],
-
1
);
PADDLE_ENFORCE_EQ
(
match_indices
[
max_idx
],
-
1
,
platform
::
errors
::
InvalidArgument
(
"The match_indices must be initialized to -1 at [%d]."
,
max_idx
));
match_indices
[
max_idx
]
=
max_row_idx
;
match_indices
[
max_idx
]
=
max_row_idx
;
match_dist
[
max_idx
]
=
max_dist
;
match_dist
[
max_idx
]
=
max_dist
;
// Erase the row index.
// Erase the row index.
...
@@ -163,7 +175,10 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
...
@@ -163,7 +175,10 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
}
}
}
}
if
(
max_row_idx
!=
-
1
)
{
if
(
max_row_idx
!=
-
1
)
{
PADDLE_ENFORCE_EQ
(
match_indices
[
j
],
-
1
);
PADDLE_ENFORCE_EQ
(
match_indices
[
j
],
-
1
,
platform
::
errors
::
InvalidArgument
(
"The match_indices must be initialized to -1 at [%d]."
,
j
));
match_indices
[
j
]
=
max_row_idx
;
match_indices
[
j
]
=
max_row_idx
;
match_dist
[
j
]
=
max_dist
;
match_dist
[
j
]
=
max_dist
;
}
}
...
@@ -183,8 +198,9 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
...
@@ -183,8 +198,9 @@ class BipartiteMatchKernel : public framework::OpKernel<T> {
?
1
?
1
:
static_cast
<
int64_t
>
(
dist_mat
->
lod
().
back
().
size
()
-
1
);
:
static_cast
<
int64_t
>
(
dist_mat
->
lod
().
back
().
size
()
-
1
);
if
(
dist_mat
->
lod
().
size
())
{
if
(
dist_mat
->
lod
().
size
())
{
PADDLE_ENFORCE_EQ
(
dist_mat
->
lod
().
size
(),
1UL
,
PADDLE_ENFORCE_EQ
(
"Only support 1 level of LoD."
);
dist_mat
->
lod
().
size
(),
1UL
,
platform
::
errors
::
InvalidArgument
(
"Only support 1 level of LoD."
));
}
}
match_indices
->
mutable_data
<
int
>
({
n
,
col
},
context
.
GetPlace
());
match_indices
->
mutable_data
<
int
>
({
n
,
col
},
context
.
GetPlace
());
match_dist
->
mutable_data
<
T
>
({
n
,
col
},
context
.
GetPlace
());
match_dist
->
mutable_data
<
T
>
({
n
,
col
},
context
.
GetPlace
());
...
...
paddle/fluid/operators/detection/generate_mask_labels_op.cc
浏览文件 @
05c9642d
...
@@ -40,35 +40,49 @@ class GenerateMaskLabelsOp : public framework::OperatorWithKernel {
...
@@ -40,35 +40,49 @@ class GenerateMaskLabelsOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"ImInfo"
),
"Input(ImInfo) shouldn't be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"GtClasses"
),
ctx
->
HasInput
(
"ImInfo"
),
true
,
"Input(GtClasses) shouldn't be null."
);
platform
::
errors
::
InvalidArgument
(
"Input(ImInfo) shouldn't be null."
));
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"IsCrowd"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"GtClasses"
),
true
,
"Input(IsCrowd) shouldn't be null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"GtSegms"
),
"Input(GtClasses) shouldn't be null."
));
"Input(GtSegms) shouldn't be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Rois"
),
"Input(Rois) shouldn't be null."
);
ctx
->
HasInput
(
"IsCrowd"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"LabelsInt32"
),
platform
::
errors
::
InvalidArgument
(
"Input(IsCrowd) shouldn't be null."
));
"Input(LabelsInt32) shouldn't be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"GtSegms"
),
true
,
PADDLE_ENFORCE
(
platform
::
errors
::
InvalidArgument
(
"Input(GtSegms) shouldn't be null."
));
ctx
->
HasOutput
(
"MaskRois"
),
PADDLE_ENFORCE_EQ
(
"Output(MaskRois) of GenerateMaskLabelsOp should not be null"
);
ctx
->
HasInput
(
"Rois"
),
true
,
PADDLE_ENFORCE
(
platform
::
errors
::
InvalidArgument
(
"Input(Rois) shouldn't be null."
));
ctx
->
HasOutput
(
"RoiHasMaskInt32"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"LabelsInt32"
),
true
,
"Output(RoiHasMaskInt32) of GenerateMaskLabelsOp should not be null"
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
"Input(LabelsInt32) shouldn't be null."
));
ctx
->
HasOutput
(
"MaskInt32"
),
"Output(MaskInt32) of GenerateMaskLabelsOp should not be null"
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"MaskRois"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(MaskRois) of GenerateMaskLabelsOp should not be null"
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"RoiHasMaskInt32"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(RoiHasMaskInt32) of GenerateMaskLabelsOp "
"should not be null"
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"MaskInt32"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(MaskInt32) of GenerateMaskLabelsOp should not be null"
));
auto
im_info_dims
=
ctx
->
GetInputDim
(
"ImInfo"
);
auto
im_info_dims
=
ctx
->
GetInputDim
(
"ImInfo"
);
auto
gt_segms_dims
=
ctx
->
GetInputDim
(
"GtSegms"
);
auto
gt_segms_dims
=
ctx
->
GetInputDim
(
"GtSegms"
);
PADDLE_ENFORCE_EQ
(
im_info_dims
.
size
(),
2
,
PADDLE_ENFORCE_EQ
(
im_info_dims
.
size
(),
2
,
"The rank of Input(ImInfo) must be 2."
);
platform
::
errors
::
InvalidArgument
(
"The rank of Input(ImInfo) must be 2."
));
PADDLE_ENFORCE_EQ
(
gt_segms_dims
.
size
(),
2
,
PADDLE_ENFORCE_EQ
(
gt_segms_dims
.
size
(),
2
,
"The rank of Input(GtSegms) must be 2."
);
platform
::
errors
::
InvalidArgument
(
"The rank of Input(GtSegms) must be 2."
));
PADDLE_ENFORCE_EQ
(
gt_segms_dims
[
1
],
2
,
PADDLE_ENFORCE_EQ
(
gt_segms_dims
[
1
],
2
,
"The second dim of Input(GtSegms) must be 2."
);
platform
::
errors
::
InvalidArgument
(
"The second dim of Input(GtSegms) must be 2."
));
int
num_classes
=
ctx
->
Attrs
().
Get
<
int
>
(
"num_classes"
);
int
num_classes
=
ctx
->
Attrs
().
Get
<
int
>
(
"num_classes"
);
int
resolution
=
ctx
->
Attrs
().
Get
<
int
>
(
"resolution"
);
int
resolution
=
ctx
->
Attrs
().
Get
<
int
>
(
"resolution"
);
...
@@ -134,7 +148,11 @@ std::vector<Tensor> SampleMaskForOneImage(
...
@@ -134,7 +148,11 @@ std::vector<Tensor> SampleMaskForOneImage(
const
int
*
gt_classes_data
=
gt_classes
.
data
<
int
>
();
const
int
*
gt_classes_data
=
gt_classes
.
data
<
int
>
();
const
int
*
is_crowd_data
=
is_crowd
.
data
<
int
>
();
const
int
*
is_crowd_data
=
is_crowd
.
data
<
int
>
();
const
int
*
label_int32_data
=
label_int32
.
data
<
int
>
();
const
int
*
label_int32_data
=
label_int32
.
data
<
int
>
();
PADDLE_ENFORCE_EQ
(
roi_size
,
label_int32
.
dims
()[
0
]);
PADDLE_ENFORCE_EQ
(
roi_size
,
label_int32
.
dims
()[
0
],
platform
::
errors
::
InvalidArgument
(
"The first dim of label [%d] is the different from "
"roi_size [%d], they should be same."
,
label_int32
.
dims
()[
0
],
roi_size
));
std
::
vector
<
int
>
mask_gt_inds
,
fg_inds
;
std
::
vector
<
int
>
mask_gt_inds
,
fg_inds
;
std
::
vector
<
std
::
vector
<
std
::
vector
<
T
>>>
gt_polys
;
std
::
vector
<
std
::
vector
<
std
::
vector
<
T
>>>
gt_polys
;
...
@@ -155,7 +173,12 @@ std::vector<Tensor> SampleMaskForOneImage(
...
@@ -155,7 +173,12 @@ std::vector<Tensor> SampleMaskForOneImage(
for
(
int
j
=
0
;
j
<
poly_num
;
++
j
)
{
for
(
int
j
=
0
;
j
<
poly_num
;
++
j
)
{
int
s
=
lod2
[
s_idx
+
j
];
int
s
=
lod2
[
s_idx
+
j
];
int
e
=
lod2
[
s_idx
+
j
+
1
];
int
e
=
lod2
[
s_idx
+
j
+
1
];
PADDLE_ENFORCE_NE
(
s
,
e
);
PADDLE_ENFORCE_NE
(
s
,
e
,
platform
::
errors
::
InvalidArgument
(
"The start point and the end point in the poly "
"segment [%d] should not be same, but received "
"the start point [%d] and the end point [%d]."
,
i
,
s
,
e
));
std
::
vector
<
T
>
plts
(
polys_data
+
s
*
2
,
polys_data
+
e
*
2
);
std
::
vector
<
T
>
plts
(
polys_data
+
s
*
2
,
polys_data
+
e
*
2
);
polys
.
push_back
(
plts
);
polys
.
push_back
(
plts
);
}
}
...
@@ -295,19 +318,34 @@ class GenerateMaskLabelsKernel : public framework::OpKernel<T> {
...
@@ -295,19 +318,34 @@ class GenerateMaskLabelsKernel : public framework::OpKernel<T> {
int
num_classes
=
ctx
.
Attr
<
int
>
(
"num_classes"
);
int
num_classes
=
ctx
.
Attr
<
int
>
(
"num_classes"
);
int
resolution
=
ctx
.
Attr
<
int
>
(
"resolution"
);
int
resolution
=
ctx
.
Attr
<
int
>
(
"resolution"
);
PADDLE_ENFORCE_EQ
(
gt_classes
->
lod
().
size
(),
1UL
,
PADDLE_ENFORCE_EQ
(
"GenerateMaskLabelsOp gt_classes needs 1 level of LoD"
);
gt_classes
->
lod
().
size
(),
1UL
,
PADDLE_ENFORCE_EQ
(
is_crowd
->
lod
().
size
(),
1UL
,
platform
::
errors
::
InvalidArgument
(
"GenerateMaskLabelsOp is_crowd needs 1 level of LoD"
);
"GenerateMaskLabelsOp gt_classes needs 1 level of LoD"
));
PADDLE_ENFORCE_EQ
(
is_crowd
->
lod
().
size
(),
1UL
,
platform
::
errors
::
InvalidArgument
(
"GenerateMaskLabelsOp is_crowd needs 1 level of LoD"
));
PADDLE_ENFORCE_EQ
(
rois
->
lod
().
size
(),
1UL
,
PADDLE_ENFORCE_EQ
(
rois
->
lod
().
size
(),
1UL
,
"GenerateMaskLabelsOp rois needs 1 level of LoD"
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_EQ
(
label_int32
->
lod
().
size
(),
1UL
,
"GenerateMaskLabelsOp rois needs 1 level of LoD"
));
"GenerateMaskLabelsOp label_int32 needs 1 level of LoD"
);
PADDLE_ENFORCE_EQ
(
label_int32
->
lod
().
size
(),
1UL
,
PADDLE_ENFORCE_EQ
(
gt_segms
->
lod
().
size
(),
3UL
);
platform
::
errors
::
InvalidArgument
(
"GenerateMaskLabelsOp label_int32 needs 1 level of LoD"
));
PADDLE_ENFORCE_EQ
(
gt_segms
->
lod
().
size
(),
3UL
,
platform
::
errors
::
InvalidArgument
(
"GenerateMaskLabelsOp gt_segms needs 3 level of LoD"
));
int64_t
n
=
static_cast
<
int64_t
>
(
gt_classes
->
lod
().
back
().
size
()
-
1
);
int64_t
n
=
static_cast
<
int64_t
>
(
gt_classes
->
lod
().
back
().
size
()
-
1
);
PADDLE_ENFORCE_EQ
(
gt_segms
->
lod
()[
0
].
size
()
-
1
,
n
);
PADDLE_ENFORCE_EQ
(
gt_segms
->
lod
()[
0
].
size
()
-
1
,
n
,
platform
::
errors
::
InvalidArgument
(
"Batchsize of Input(gt_segms) and Input(gt_classes) should be "
"same, but received gt_segms[%d], gt_classes[%d]."
,
gt_segms
->
lod
()[
0
].
size
()
-
1
,
n
));
int
mask_dim
=
num_classes
*
resolution
*
resolution
;
int
mask_dim
=
num_classes
*
resolution
*
resolution
;
int
roi_num
=
rois
->
lod
().
back
()[
n
];
int
roi_num
=
rois
->
lod
().
back
()[
n
];
...
...
paddle/fluid/operators/detection/target_assign_op.cc
浏览文件 @
05c9642d
...
@@ -22,29 +22,41 @@ class TargetAssignOp : public framework::OperatorWithKernel {
...
@@ -22,29 +22,41 @@ class TargetAssignOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
"Input(X) of TargetAssignOp should not be null"
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"MatchIndices"
),
"Input(X) of TargetAssignOp should not be null"
));
"Input(MatchIndices) of TargetAssignOp should not be null"
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"MatchIndices"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
platform
::
errors
::
InvalidArgument
(
"Output(Out) of TargetAssignOp should not be null."
);
"Input(MatchIndices) of TargetAssignOp should not be null"
));
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"OutWeight"
),
"Output(OutWeight) of TargetAssignOp should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Out) of TargetAssignOp should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"OutWeight"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(OutWeight) of TargetAssignOp should not be null."
));
auto
in_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
in_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
mi_dims
=
ctx
->
GetInputDim
(
"MatchIndices"
);
auto
mi_dims
=
ctx
->
GetInputDim
(
"MatchIndices"
);
PADDLE_ENFORCE_EQ
(
in_dims
.
size
(),
3
,
"The rank of Input(X) must be 3."
);
PADDLE_ENFORCE_EQ
(
in_dims
.
size
(),
3
,
platform
::
errors
::
InvalidArgument
(
"The rank of Input(X) must be 3."
));
PADDLE_ENFORCE_EQ
(
mi_dims
.
size
(),
2
,
PADDLE_ENFORCE_EQ
(
mi_dims
.
size
(),
2
,
"The rank of Input(MatchIndices) must be 2."
);
platform
::
errors
::
InvalidArgument
(
"The rank of Input(MatchIndices) must be 2."
));
if
(
ctx
->
HasInput
(
"NegIndices"
))
{
if
(
ctx
->
HasInput
(
"NegIndices"
))
{
auto
neg_dims
=
ctx
->
GetInputDim
(
"NegIndices"
);
auto
neg_dims
=
ctx
->
GetInputDim
(
"NegIndices"
);
PADDLE_ENFORCE_EQ
(
neg_dims
.
size
(),
2
,
PADDLE_ENFORCE_EQ
(
neg_dims
.
size
(),
2
,
"The rank of Input(NegIndices) must be 2."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_EQ
(
neg_dims
[
1
],
1
,
"The rank of Input(NegIndices) must be 2."
));
"The last dimension of Out(NegIndices) must be 1."
);
PADDLE_ENFORCE_EQ
(
neg_dims
[
1
],
1
,
platform
::
errors
::
InvalidArgument
(
"The last dimension of Out(NegIndices) must be 1."
));
}
}
auto
n
=
mi_dims
[
0
];
auto
n
=
mi_dims
[
0
];
...
...
paddle/fluid/operators/detection/target_assign_op.h
浏览文件 @
05c9642d
...
@@ -90,7 +90,9 @@ class TargetAssignKernel : public framework::OpKernel<T> {
...
@@ -90,7 +90,9 @@ class TargetAssignKernel : public framework::OpKernel<T> {
auto
*
out
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
auto
*
out
=
ctx
.
Output
<
framework
::
Tensor
>
(
"Out"
);
auto
*
out_wt
=
ctx
.
Output
<
framework
::
Tensor
>
(
"OutWeight"
);
auto
*
out_wt
=
ctx
.
Output
<
framework
::
Tensor
>
(
"OutWeight"
);
PADDLE_ENFORCE_EQ
(
x
->
lod
().
size
(),
1UL
);
PADDLE_ENFORCE_EQ
(
x
->
lod
().
size
(),
1UL
,
platform
::
errors
::
InvalidArgument
(
"TargetAssignOp input(X) needs 1 level of LoD"
));
int
mismatch_value
=
ctx
.
Attr
<
int
>
(
"mismatch_value"
);
int
mismatch_value
=
ctx
.
Attr
<
int
>
(
"mismatch_value"
);
const
T
*
x_data
=
x
->
data
<
T
>
();
const
T
*
x_data
=
x
->
data
<
T
>
();
...
@@ -121,7 +123,10 @@ class TargetAssignKernel : public framework::OpKernel<T> {
...
@@ -121,7 +123,10 @@ class TargetAssignKernel : public framework::OpKernel<T> {
auto
*
neg_indices
=
ctx
.
Input
<
framework
::
LoDTensor
>
(
"NegIndices"
);
auto
*
neg_indices
=
ctx
.
Input
<
framework
::
LoDTensor
>
(
"NegIndices"
);
if
(
neg_indices
)
{
if
(
neg_indices
)
{
PADDLE_ENFORCE_EQ
(
neg_indices
->
lod
().
size
(),
1UL
);
PADDLE_ENFORCE_EQ
(
neg_indices
->
lod
().
size
(),
1UL
,
platform
::
errors
::
InvalidArgument
(
"TargetAssignOp input(NegIndices) needs 1 level of LoD"
));
const
int
*
neg_idx_data
=
neg_indices
->
data
<
int
>
();
const
int
*
neg_idx_data
=
neg_indices
->
data
<
int
>
();
auto
neg_lod
=
neg_indices
->
lod
().
back
();
auto
neg_lod
=
neg_indices
->
lod
().
back
();
#if defined(PADDLE_WITH_CUDA)
#if defined(PADDLE_WITH_CUDA)
...
...
paddle/fluid/operators/filter_by_instag_op.cc
浏览文件 @
05c9642d
...
@@ -24,19 +24,25 @@ class FilterByInstagOp : public framework::OperatorWithKernel {
...
@@ -24,19 +24,25 @@ class FilterByInstagOp : public framework::OperatorWithKernel {
public:
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Ins"
),
true
,
PADDLE_ENFORCE_EQ
(
"Input(Ins) should be not null."
);
ctx
->
HasInput
(
"Ins"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Ins) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Ins_tag"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Ins_tag"
),
true
,
"Input(Ins_tag) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Ins_tag) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Filter_tag"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Filter_tag"
),
true
,
"Input(Filter_tag) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Filter_tag) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
PADDLE_ENFORCE_EQ
(
"Output(Out) should be not null."
);
ctx
->
HasOutput
(
"Out"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Out) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"LossWeight"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"LossWeight"
),
true
,
"Output(LossWeight) shoudl not be null."
);
platform
::
errors
::
InvalidArgument
(
"Output(LossWeight) shoudl not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"IndexMap"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"IndexMap"
),
true
,
"Output(IndexMap) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Output(IndexMap) should be not null."
));
auto
x1_dims
=
ctx
->
GetInputDim
(
"Ins"
);
// batch_size * vec
auto
x1_dims
=
ctx
->
GetInputDim
(
"Ins"
);
// batch_size * vec
...
@@ -85,15 +91,20 @@ class FilterByInstagOpGrad : public framework::OperatorWithKernel {
...
@@ -85,15 +91,20 @@ class FilterByInstagOpGrad : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"IndexMap"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"IndexMap"
),
true
,
"Input(IndexMap) should be not null"
);
platform
::
errors
::
InvalidArgument
(
"Input(IndexMap) should be not null"
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
"Grad Input(Out) should be not null"
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Ins"
),
true
,
"Grad Input(Out) should be not null"
));
"Input(Ins) should be not null"
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Ins"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Ins) should be not null"
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"LossWeight"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"LossWeight"
),
true
,
"Input(LossWeight) should be not null"
);
platform
::
errors
::
InvalidArgument
(
"Input(LossWeight) should be not null"
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"Ins"
)),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"Ins"
)),
true
,
"Grad Output(Ins) should be not null"
);
platform
::
errors
::
InvalidArgument
(
"Grad Output(Ins) should be not null"
));
auto
grad_out_dims
=
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
));
auto
grad_out_dims
=
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
));
auto
x1_dims
=
ctx
->
GetInputDim
(
"Ins"
);
auto
x1_dims
=
ctx
->
GetInputDim
(
"Ins"
);
...
...
paddle/fluid/operators/one_hot_v2_op.h
浏览文件 @
05c9642d
...
@@ -51,11 +51,19 @@ struct OneHotV2OpFunctor {
...
@@ -51,11 +51,19 @@ struct OneHotV2OpFunctor {
}
}
}
else
{
}
else
{
for
(
int
i
=
0
;
i
<
numel
;
++
i
)
{
for
(
int
i
=
0
;
i
<
numel
;
++
i
)
{
PADDLE_ENFORCE_GE
(
p_in_data
[
i
],
0
,
PADDLE_ENFORCE_GE
(
"Illegal index value, should be at least 0."
);
p_in_data
[
i
],
0
,
platform
::
errors
::
InvalidArgument
(
"Illegal index value, Input(input) value should be at least 0, "
"but received input (%d) less than 0"
,
p_in_data
[
i
]));
PADDLE_ENFORCE_LT
(
PADDLE_ENFORCE_LT
(
p_in_data
[
i
],
depth_
,
p_in_data
[
i
],
depth_
,
"Illegal index value, should be less than depth (%d)."
,
depth_
);
platform
::
errors
::
InvalidArgument
(
"Illegal index value, Input(input) value should be less than "
"Input(depth), "
"but received input (%d) not less than depth (%d)"
,
p_in_data
[
i
],
depth_
));
*
(
p_out_data
+
i
*
depth_
+
p_in_data
[
i
])
=
1.0
;
*
(
p_out_data
+
i
*
depth_
+
p_in_data
[
i
])
=
1.0
;
}
}
}
}
...
...
paddle/fluid/operators/pool_with_index_op.cc
浏览文件 @
05c9642d
...
@@ -29,12 +29,15 @@ class MaxPoolWithIndexOp : public framework::OperatorWithKernel {
...
@@ -29,12 +29,15 @@ class MaxPoolWithIndexOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
"Input(X) of Pooling should not be null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
"Input(X) of Pooling should not be null."
));
"Output(Out) of Pooling should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Mask"
),
platform
::
errors
::
InvalidArgument
(
"Output(Mask) of Pooling should not be null."
);
"Output(Out) of Pooling should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Mask"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Mask) of Pooling should not be null."
));
auto
in_x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
in_x_dims
=
ctx
->
GetInputDim
(
"X"
);
...
@@ -54,12 +57,16 @@ class MaxPoolWithIndexOp : public framework::OperatorWithKernel {
...
@@ -54,12 +57,16 @@ class MaxPoolWithIndexOp : public framework::OperatorWithKernel {
}
}
}
}
PADDLE_ENFORCE
(
in_x_dims
.
size
()
-
ksize
.
size
()
==
2U
,
PADDLE_ENFORCE_EQ
(
in_x_dims
.
size
()
-
ksize
.
size
(),
2U
,
"Input size and pooling size should be consistent."
);
platform
::
errors
::
InvalidArgument
(
"Input size and pooling size should be consistent."
));
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
strides
.
size
(),
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
strides
.
size
(),
"Strides size and pooling size should be the same."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
paddings
.
size
(),
"Strides size and pooling size should be the same."
));
"Paddings size and pooling size should be the same."
);
PADDLE_ENFORCE_EQ
(
ksize
.
size
(),
paddings
.
size
(),
platform
::
errors
::
InvalidArgument
(
"Paddings size and pooling size should be the same."
));
std
::
vector
<
int64_t
>
output_shape
({
in_x_dims
[
0
],
in_x_dims
[
1
]});
std
::
vector
<
int64_t
>
output_shape
({
in_x_dims
[
0
],
in_x_dims
[
1
]});
if
(
adaptive
)
{
if
(
adaptive
)
{
...
@@ -90,15 +97,16 @@ class MaxPoolWithIndexOpGrad : public framework::OperatorWithKernel {
...
@@ -90,15 +97,16 @@ class MaxPoolWithIndexOpGrad : public framework::OperatorWithKernel {
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Mask"
),
true
,
ctx
->
HasInput
(
"Mask"
),
true
,
platform
::
errors
::
NotFound
(
"Input(Mask) must not be null."
));
platform
::
errors
::
InvalidArgument
(
"Input(Mask) must not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
platform
::
errors
::
NotFound
(
"Input(X) must not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
platform
::
errors
::
NotFound
(
"Input(Out@GRAD) should not be null."
));
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
true
,
ctx
->
HasInput
(
"X"
),
true
,
platform
::
errors
::
NotFound
(
"Output(X@GRAD) should not be null."
));
platform
::
errors
::
InvalidArgument
(
"Input(X) must not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Out@GRAD) should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(X@GRAD) should not be null."
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
}
}
...
...
paddle/fluid/operators/psroi_pool_op.cc
浏览文件 @
05c9642d
...
@@ -81,43 +81,57 @@ class PSROIPoolOp : public framework::OperatorWithKernel {
...
@@ -81,43 +81,57 @@ class PSROIPoolOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
"Input(X) of PSROIPoolOp should not be null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"ROIs"
),
"Input(X) of PSROIPoolOp should not be null."
));
"Input(ROIs) of PSROIPoolOp should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"ROIs"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
platform
::
errors
::
InvalidArgument
(
"Output(Out) of PSROIPoolOp should not be null."
);
"Input(ROIs) of PSROIPoolOp should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Out) of PSROIPoolOp should not be null."
));
auto
input_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
input_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
rois_dims
=
ctx
->
GetInputDim
(
"ROIs"
);
auto
rois_dims
=
ctx
->
GetInputDim
(
"ROIs"
);
PADDLE_ENFORCE
(
input_dims
.
size
()
==
4
,
PADDLE_ENFORCE_EQ
(
input_dims
.
size
(),
4
,
"The format of input tensor is NCHW"
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
rois_dims
.
size
()
==
2
,
"The format of input tensor is NCHW"
));
"ROIs should be a 2-D LoDTensor of shape (num_rois, 4) "
PADDLE_ENFORCE_EQ
(
"given as [(x1, y1, x2, y2), ...]"
);
rois_dims
.
size
(),
2
,
PADDLE_ENFORCE
(
rois_dims
[
1
]
==
4
,
platform
::
errors
::
InvalidArgument
(
"ROIs should be a 2-D LoDTensor of shape (num_rois, 4) "
"ROIs should be a 2-D LoDTensor of shape (num_rois, 4) "
"given as [(x1, y1, x2, y2), ...]"
);
"given as [(x1, y1, x2, y2), ...]"
));
PADDLE_ENFORCE_EQ
(
rois_dims
[
1
],
4
,
platform
::
errors
::
InvalidArgument
(
"ROIs should be a 2-D LoDTensor of shape (num_rois, 4) "
"given as [(x1, y1, x2, y2), ...]"
));
int
pooled_height
=
ctx
->
Attrs
().
Get
<
int
>
(
"pooled_height"
);
int
pooled_height
=
ctx
->
Attrs
().
Get
<
int
>
(
"pooled_height"
);
int
pooled_width
=
ctx
->
Attrs
().
Get
<
int
>
(
"pooled_width"
);
int
pooled_width
=
ctx
->
Attrs
().
Get
<
int
>
(
"pooled_width"
);
int
output_channels
=
ctx
->
Attrs
().
Get
<
int
>
(
"output_channels"
);
int
output_channels
=
ctx
->
Attrs
().
Get
<
int
>
(
"output_channels"
);
float
spatial_scale
=
ctx
->
Attrs
().
Get
<
float
>
(
"spatial_scale"
);
float
spatial_scale
=
ctx
->
Attrs
().
Get
<
float
>
(
"spatial_scale"
);
PADDLE_ENFORCE
(
PADDLE_ENFORCE_EQ
(
input_dims
[
1
]
==
output_channels
*
pooled_height
*
pooled_width
,
input_dims
[
1
],
output_channels
*
pooled_height
*
pooled_width
,
"the channel of X(%d) should be equal to the product of "
platform
::
errors
::
InvalidArgument
(
"output_channels(%d), pooled_height(%d) and pooled_width(%d)"
,
"the channel of X(%d) "
input_dims
[
1
],
output_channels
,
pooled_height
,
pooled_width
);
"should be equal to the product of "
"output_channels(%d), pooled_height(%d) and pooled_width(%d)"
,
input_dims
[
1
],
output_channels
,
pooled_height
,
pooled_width
));
PADDLE_ENFORCE_GT
(
pooled_height
,
0
,
PADDLE_ENFORCE_GT
(
pooled_height
,
0
,
"The pooled output height must be greater than 0"
);
platform
::
errors
::
InvalidArgument
(
"The pooled output height must be greater than 0"
));
PADDLE_ENFORCE_GT
(
pooled_width
,
0
,
PADDLE_ENFORCE_GT
(
pooled_width
,
0
,
"The pooled output width must be greater than 0"
);
platform
::
errors
::
InvalidArgument
(
"The pooled output width must be greater than 0"
));
PADDLE_ENFORCE_GT
(
output_channels
,
1
,
PADDLE_ENFORCE_GT
(
output_channels
,
1
,
"The pooled output channels must greater than 1"
);
platform
::
errors
::
InvalidArgument
(
"The pooled output channels must greater than 1"
));
PADDLE_ENFORCE_GT
(
spatial_scale
,
0.0
f
,
PADDLE_ENFORCE_GT
(
spatial_scale
,
0.0
f
,
"The spatial scale must greater than 0."
);
platform
::
errors
::
InvalidArgument
(
"The spatial scale must greater than 0."
));
auto
out_dims
=
input_dims
;
auto
out_dims
=
input_dims
;
out_dims
[
0
]
=
rois_dims
[
0
];
out_dims
[
0
]
=
rois_dims
[
0
];
...
@@ -142,10 +156,12 @@ class PSROIPoolGradOp : public framework::OperatorWithKernel {
...
@@ -142,10 +156,12 @@ class PSROIPoolGradOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
"The gradient of Out should not be null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
"The gradient of Out should not be null."
));
"The gradient of X should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"The gradient of X should not be null."
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
}
}
...
...
paddle/fluid/operators/psroi_pool_op.h
浏览文件 @
05c9642d
...
@@ -54,15 +54,19 @@ class CPUPSROIPoolOpKernel : public framework::OpKernel<T> {
...
@@ -54,15 +54,19 @@ class CPUPSROIPoolOpKernel : public framework::OpKernel<T> {
int
rois_batch_size
=
rois_lod
.
size
()
-
1
;
int
rois_batch_size
=
rois_lod
.
size
()
-
1
;
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
rois_batch_size
,
batch_size
,
rois_batch_size
,
batch_size
,
"the rois_batch_size and input(X) batch_size should be the same."
);
platform
::
errors
::
InvalidArgument
(
"the rois_batch_size and input(X) "
"batch_size should be the same."
));
int
rois_num_with_lod
=
rois_lod
[
rois_batch_size
];
int
rois_num_with_lod
=
rois_lod
[
rois_batch_size
];
PADDLE_ENFORCE_EQ
(
rois_num_with_lod
,
rois_num
,
PADDLE_ENFORCE_EQ
(
rois_num_with_lod
,
rois_num
,
"the rois_num from input and lod must be the same"
);
platform
::
errors
::
InvalidArgument
(
"the rois_num from input and lod must be the same"
));
PADDLE_ENFORCE_EQ
(
input_channels
,
PADDLE_ENFORCE_EQ
(
input_channels
,
output_channels
*
pooled_height
*
pooled_width
,
output_channels
*
pooled_height
*
pooled_width
,
"the channels of input X should equal the product of "
platform
::
errors
::
InvalidArgument
(
"output_channels x pooled_height x pooled_width"
);
"the channels of input "
"X should equal the product of "
"output_channels x pooled_height x pooled_width"
));
// calculate batch id index for each roi according to LoD
// calculate batch id index for each roi according to LoD
for
(
int
n
=
0
;
n
<
rois_batch_size
;
++
n
)
{
for
(
int
n
=
0
;
n
<
rois_batch_size
;
++
n
)
{
...
...
paddle/fluid/operators/roi_pool_op.cc
浏览文件 @
05c9642d
...
@@ -36,7 +36,10 @@ class ROIPoolOp : public framework::OperatorWithKernel {
...
@@ -36,7 +36,10 @@ class ROIPoolOp : public framework::OperatorWithKernel {
if
(
ctx
->
HasInput
(
"RoisLod"
))
{
if
(
ctx
->
HasInput
(
"RoisLod"
))
{
auto
rois_lod_dims
=
ctx
->
GetInputDim
(
"RoisLod"
);
auto
rois_lod_dims
=
ctx
->
GetInputDim
(
"RoisLod"
);
PADDLE_ENFORCE
(
rois_lod_dims
.
size
()
==
1
,
""
);
PADDLE_ENFORCE_EQ
(
rois_lod_dims
.
size
(),
1
,
platform
::
errors
::
InvalidArgument
(
"The lod information tensor of ROIs should "
"be one-dimensional"
));
}
}
PADDLE_ENFORCE_EQ
(
input_dims
.
size
(),
4
,
PADDLE_ENFORCE_EQ
(
input_dims
.
size
(),
4
,
platform
::
errors
::
InvalidArgument
(
platform
::
errors
::
InvalidArgument
(
...
...
paddle/fluid/operators/roi_pool_op.h
浏览文件 @
05c9642d
...
@@ -63,7 +63,8 @@ class CPUROIPoolOpKernel : public framework::OpKernel<T> {
...
@@ -63,7 +63,8 @@ class CPUROIPoolOpKernel : public framework::OpKernel<T> {
rois_batch_size
=
rois_lod_t
->
numel
();
rois_batch_size
=
rois_lod_t
->
numel
();
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
rois_batch_size
-
1
,
batch_size
,
rois_batch_size
-
1
,
batch_size
,
"The rois_batch_size and imgs batch_size must be the same."
);
platform
::
errors
::
InvalidArgument
(
"The rois_batch_size and imgs "
"batch_size must be the same."
));
auto
*
rois_lod
=
rois_lod_t
->
data
<
int64_t
>
();
auto
*
rois_lod
=
rois_lod_t
->
data
<
int64_t
>
();
for
(
int
n
=
0
;
n
<
rois_batch_size
-
1
;
++
n
)
{
for
(
int
n
=
0
;
n
<
rois_batch_size
-
1
;
++
n
)
{
for
(
int
i
=
rois_lod
[
n
];
i
<
rois_lod
[
n
+
1
];
++
i
)
{
for
(
int
i
=
rois_lod
[
n
];
i
<
rois_lod
[
n
+
1
];
++
i
)
{
...
@@ -75,10 +76,13 @@ class CPUROIPoolOpKernel : public framework::OpKernel<T> {
...
@@ -75,10 +76,13 @@ class CPUROIPoolOpKernel : public framework::OpKernel<T> {
rois_batch_size
=
rois_lod
.
size
()
-
1
;
rois_batch_size
=
rois_lod
.
size
()
-
1
;
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
rois_batch_size
,
batch_size
,
rois_batch_size
,
batch_size
,
"The rois_batch_size and imgs batch_size must be the same."
);
platform
::
errors
::
InvalidArgument
(
"The rois_batch_size and imgs "
"batch_size must be the same."
));
int
rois_num_with_lod
=
rois_lod
[
rois_batch_size
];
int
rois_num_with_lod
=
rois_lod
[
rois_batch_size
];
PADDLE_ENFORCE_EQ
(
rois_num
,
rois_num_with_lod
,
PADDLE_ENFORCE_EQ
(
"The rois_num from input and lod must be the same."
);
rois_num
,
rois_num_with_lod
,
platform
::
errors
::
InvalidArgument
(
"The rois_num from input "
"and lod must be the same."
));
for
(
int
n
=
0
;
n
<
rois_batch_size
;
++
n
)
{
for
(
int
n
=
0
;
n
<
rois_batch_size
;
++
n
)
{
for
(
size_t
i
=
rois_lod
[
n
];
i
<
rois_lod
[
n
+
1
];
++
i
)
{
for
(
size_t
i
=
rois_lod
[
n
];
i
<
rois_lod
[
n
+
1
];
++
i
)
{
roi_batch_id_data
[
i
]
=
n
;
roi_batch_id_data
[
i
]
=
n
;
...
...
paddle/fluid/operators/softmax_op.cc
浏览文件 @
05c9642d
...
@@ -34,21 +34,30 @@ class SoftmaxOp : public framework::OperatorWithKernel {
...
@@ -34,21 +34,30 @@ class SoftmaxOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
PADDLE_ENFORCE_EQ
(
"Input(X) of SoftmaxOp should not be null."
);
ctx
->
HasInput
(
"X"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
platform
::
errors
::
NotFound
(
"Input(X) of SoftmaxOp is not found."
));
"Output(Out) of SoftmaxOp should not be null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
platform
::
errors
::
NotFound
(
"Output(Out) of SoftmaxOp is not found."
));
auto
dim_x
=
ctx
->
GetInputDim
(
"X"
);
auto
dim_x
=
ctx
->
GetInputDim
(
"X"
);
auto
rank_x
=
dim_x
.
size
();
auto
rank_x
=
dim_x
.
size
();
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
PADDLE_ENFORCE
(
axis
>=
-
rank_x
&&
axis
<
rank_x
,
PADDLE_ENFORCE_GE
(
axis
,
-
rank_x
,
"Attr(axis) value should be in range [-R, R-1], "
platform
::
errors
::
InvalidArgument
(
"R is the rank of Input(X)."
);
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(X)."
));
PADDLE_ENFORCE_LT
(
axis
,
rank_x
,
platform
::
errors
::
InvalidArgument
(
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(X)."
));
auto
use_cudnn
=
ctx
->
Attrs
().
Get
<
bool
>
(
"use_cudnn"
);
auto
use_cudnn
=
ctx
->
Attrs
().
Get
<
bool
>
(
"use_cudnn"
);
if
(
axis
!=
rank_x
-
1
&&
axis
!=
-
1
)
{
if
(
axis
!=
rank_x
-
1
&&
axis
!=
-
1
)
{
PADDLE_ENFORCE
(
!
use_cudnn
,
"CUDNN kernel only support axis as -1."
);
PADDLE_ENFORCE_EQ
(
use_cudnn
,
false
,
platform
::
errors
::
InvalidArgument
(
"CUDNN kernel only support axis as -1."
));
}
}
ctx
->
SetOutputDim
(
"Out"
,
ctx
->
GetInputDim
(
"X"
));
ctx
->
SetOutputDim
(
"Out"
,
ctx
->
GetInputDim
(
"X"
));
...
@@ -78,8 +87,9 @@ class SoftmaxOp : public framework::OperatorWithKernel {
...
@@ -78,8 +87,9 @@ class SoftmaxOp : public framework::OperatorWithKernel {
auto
input_data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
auto
input_data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
"X"
);
if
(
input_data_type
==
framework
::
proto
::
VarType
::
FP16
)
{
if
(
input_data_type
==
framework
::
proto
::
VarType
::
FP16
)
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
PADDLE_ENFORCE_EQ
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
true
,
"float16 can only be used on GPU place"
);
platform
::
errors
::
InvalidArgument
(
"float16 can only be used on GPU place"
));
}
}
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
...
@@ -157,12 +167,17 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
...
@@ -157,12 +167,17 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Out"
),
"Input(Out) should be not null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
ctx
->
HasInput
(
"Out"
),
true
,
"Input(Out@GRAD) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Out) is not found."
));
PADDLE_ENFORCE_EQ
(
ctx
->
GetInputDim
(
"Out"
),
PADDLE_ENFORCE_EQ
(
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
)),
ctx
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
"Input(Out) and its gradients should have a same shape."
);
platform
::
errors
::
InvalidArgument
(
"Input(Out@GRAD) is not found."
));
PADDLE_ENFORCE_EQ
(
ctx
->
GetInputDim
(
"Out"
),
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
)),
platform
::
errors
::
InvalidArgument
(
"Input(Out) and its gradients "
"should have a same shape."
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
)));
ctx
->
GetInputDim
(
framework
::
GradVarName
(
"Out"
)));
...
@@ -191,8 +206,9 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
...
@@ -191,8 +206,9 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel {
auto
input_data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
auto
input_data_type
=
OperatorWithKernel
::
IndicateVarDataType
(
ctx
,
framework
::
GradVarName
(
"Out"
));
ctx
,
framework
::
GradVarName
(
"Out"
));
if
(
input_data_type
==
framework
::
proto
::
VarType
::
FP16
)
{
if
(
input_data_type
==
framework
::
proto
::
VarType
::
FP16
)
{
PADDLE_ENFORCE
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
PADDLE_ENFORCE_EQ
(
platform
::
is_gpu_place
(
ctx
.
GetPlace
()),
true
,
"float16 can only be used on GPU place"
);
platform
::
errors
::
InvalidArgument
(
"float16 can only be used on GPU place"
));
}
}
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
return
framework
::
OpKernelType
(
input_data_type
,
ctx
.
GetPlace
(),
layout_
,
...
...
paddle/fluid/operators/softmax_with_cross_entropy_op.cc
浏览文件 @
05c9642d
...
@@ -108,39 +108,51 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
...
@@ -108,39 +108,51 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Logits"
),
PADDLE_ENFORCE_EQ
(
"Input(Logits) should be not null."
);
ctx
->
HasInput
(
"Logits"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Label"
),
"Input(Label) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Logits) should be not null."
));
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Softmax"
),
ctx
->
HasInput
(
"Label"
),
true
,
"Output(Softmax) should be not null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Label) should be not null."
));
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Loss"
),
"Output(Loss) should be not null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Softmax"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Softmax) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Loss"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Loss) should be not null."
));
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
auto
logits_dims
=
ctx
->
GetInputDim
(
"Logits"
);
auto
logits_dims
=
ctx
->
GetInputDim
(
"Logits"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
auto
logits_rank
=
logits_dims
.
size
();
auto
logits_rank
=
logits_dims
.
size
();
PADDLE_ENFORCE
(
axis
>=
-
logits_rank
&&
axis
<
logits_rank
,
PADDLE_ENFORCE_GE
(
axis
,
-
logits_rank
,
"Attr(axis) value should be in range [-R, R-1], "
platform
::
errors
::
InvalidArgument
(
"R is the rank of Input(Logits)."
);
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(Logits)."
));
PADDLE_ENFORCE_LT
(
axis
,
logits_rank
,
platform
::
errors
::
InvalidArgument
(
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(Logits)."
));
axis
=
CanonicalAxis
(
axis
,
logits_rank
);
axis
=
CanonicalAxis
(
axis
,
logits_rank
);
for
(
int
i
=
0
;
i
<
logits_rank
;
i
++
)
{
for
(
int
i
=
0
;
i
<
logits_rank
;
i
++
)
{
if
(
i
!=
axis
)
{
if
(
i
!=
axis
)
{
if
(
ctx
->
IsRuntime
()
||
(
logits_dims
[
i
]
>
0
&&
labels_dims
[
i
]
>
0
))
{
if
(
ctx
->
IsRuntime
()
||
(
logits_dims
[
i
]
>
0
&&
labels_dims
[
i
]
>
0
))
{
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
logits_dims
[
i
],
labels_dims
[
i
],
logits_dims
[
i
],
labels_dims
[
i
],
platform
::
errors
::
InvalidArgument
(
"Input(Logits) and Input(Label) should in same shape
in "
"Input(Logits) and Input(Label) should
in "
"dimensions except axis."
);
"same shape in dimensions except axis."
)
);
}
}
}
}
}
}
auto
numeric_stable_mode
=
ctx
->
Attrs
().
Get
<
bool
>
(
"numeric_stable_mode"
);
auto
numeric_stable_mode
=
ctx
->
Attrs
().
Get
<
bool
>
(
"numeric_stable_mode"
);
if
(
axis
!=
logits_rank
-
1
)
{
if
(
axis
!=
logits_rank
-
1
)
{
PADDLE_ENFORCE
(
PADDLE_ENFORCE_EQ
(
numeric_stable_mode
,
true
,
numeric_stable_mode
,
platform
::
errors
::
InvalidArgument
(
"Attr(axis) can only be -1 when not in numeric_stable_mode."
);
"Attr(axis) can only be -1 "
"when not in numeric_stable_mode."
));
}
}
bool
soft_label
=
ctx
->
Attrs
().
Get
<
bool
>
(
"soft_label"
);
bool
soft_label
=
ctx
->
Attrs
().
Get
<
bool
>
(
"soft_label"
);
...
@@ -148,14 +160,18 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
...
@@ -148,14 +160,18 @@ class SoftmaxWithCrossEntropyOp : public framework::OperatorWithKernel {
if
(
ctx
->
IsRuntime
()
||
if
(
ctx
->
IsRuntime
()
||
(
logits_dims
[
axis
]
>
0
&&
labels_dims
[
axis
]
>
0
))
{
(
logits_dims
[
axis
]
>
0
&&
labels_dims
[
axis
]
>
0
))
{
PADDLE_ENFORCE_EQ
(
logits_dims
[
axis
],
labels_dims
[
axis
],
PADDLE_ENFORCE_EQ
(
logits_dims
[
axis
],
labels_dims
[
axis
],
"If Attr(soft_label) == true, the axis dimension of "
platform
::
errors
::
InvalidArgument
(
"Input(X) and Input(Label) should be equal."
);
"If Attr(soft_label) == true, "
"the axis dimension of "
"Input(X) and Input(Label) should be equal."
));
}
}
}
else
{
}
else
{
if
(
ctx
->
IsRuntime
()
||
labels_dims
[
axis
]
>
0
)
{
if
(
ctx
->
IsRuntime
()
||
labels_dims
[
axis
]
>
0
)
{
PADDLE_ENFORCE_EQ
(
labels_dims
[
axis
],
1UL
,
PADDLE_ENFORCE_EQ
(
"If Attr(soft_label) == false, the axis dimension of "
labels_dims
[
axis
],
1UL
,
"Input(Label) should be 1."
);
platform
::
errors
::
InvalidArgument
(
"If Attr(soft_label) == false, "
"the axis dimension of "
"Input(Label) should be 1."
));
}
}
}
}
...
@@ -182,21 +198,31 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
...
@@ -182,21 +198,31 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Loss"
)),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
framework
::
GradVarName
(
"Loss"
)),
true
,
"Input(Loss@Grad) should not be null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Softmax"
),
"Input(Loss@Grad) should not be null."
));
"Input(Softmax) should be not null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Softmax"
),
true
,
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"Label"
),
"Input(Label) should be not null."
);
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"Logits"
)),
"Input(Softmax) should be not null."
));
"Output(Logits@Grad) should be not null."
);
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"Label"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(Label) should be not null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"Logits"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"Output(Logits@Grad) should be not null."
));
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
auto
axis
=
ctx
->
Attrs
().
Get
<
int
>
(
"axis"
);
auto
softmax_dims
=
ctx
->
GetInputDim
(
"Softmax"
);
auto
softmax_dims
=
ctx
->
GetInputDim
(
"Softmax"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
auto
labels_dims
=
ctx
->
GetInputDim
(
"Label"
);
auto
softmax_rank
=
softmax_dims
.
size
();
auto
softmax_rank
=
softmax_dims
.
size
();
PADDLE_ENFORCE
(
axis
>=
-
softmax_rank
&&
axis
<
softmax_rank
,
PADDLE_ENFORCE_GE
(
axis
,
-
softmax_rank
,
"Attr(axis) value should be in range [-R, R-1], "
platform
::
errors
::
InvalidArgument
(
"R is the rank of Input(Logits)."
);
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(Logits)."
));
PADDLE_ENFORCE_LT
(
axis
,
softmax_rank
,
platform
::
errors
::
InvalidArgument
(
"Attr(axis) value should be in range [-R, R-1], "
"R is the rank of Input(Logits)."
));
axis
=
CanonicalAxis
(
axis
,
softmax_rank
);
axis
=
CanonicalAxis
(
axis
,
softmax_rank
);
for
(
int
i
=
0
;
i
<
softmax_rank
;
i
++
)
{
for
(
int
i
=
0
;
i
<
softmax_rank
;
i
++
)
{
...
@@ -204,8 +230,9 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
...
@@ -204,8 +230,9 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
if
(
ctx
->
IsRuntime
()
||
(
softmax_dims
[
i
]
>
0
&&
labels_dims
[
i
]
>
0
))
{
if
(
ctx
->
IsRuntime
()
||
(
softmax_dims
[
i
]
>
0
&&
labels_dims
[
i
]
>
0
))
{
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
softmax_dims
[
i
],
labels_dims
[
i
],
softmax_dims
[
i
],
labels_dims
[
i
],
"Input(Logits) and Input(Label) should in same shape in "
platform
::
errors
::
InvalidArgument
(
"dimensions except axis."
);
"Input(Logits) and Input(Label) should in same shape in "
"dimensions except axis."
));
}
}
}
}
}
}
...
@@ -215,14 +242,18 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
...
@@ -215,14 +242,18 @@ class SoftmaxWithCrossEntropyOpGrad : public framework::OperatorWithKernel {
if
(
ctx
->
IsRuntime
()
||
if
(
ctx
->
IsRuntime
()
||
(
softmax_dims
[
axis
]
>
0
&&
labels_dims
[
axis
]
>
0
))
{
(
softmax_dims
[
axis
]
>
0
&&
labels_dims
[
axis
]
>
0
))
{
PADDLE_ENFORCE_EQ
(
softmax_dims
[
axis
],
labels_dims
[
axis
],
PADDLE_ENFORCE_EQ
(
softmax_dims
[
axis
],
labels_dims
[
axis
],
"If Attr(soft_label) == true, the axis dimension of "
platform
::
errors
::
InvalidArgument
(
"Input(X) and Input(Label) should be equal."
);
"If Attr(soft_label) == true, "
"the axis dimension of "
"Input(X) and Input(Label) should be equal."
));
}
}
}
else
{
}
else
{
if
(
ctx
->
IsRuntime
()
||
labels_dims
[
axis
]
>
0
)
{
if
(
ctx
->
IsRuntime
()
||
labels_dims
[
axis
]
>
0
)
{
PADDLE_ENFORCE_EQ
(
labels_dims
[
axis
],
1UL
,
PADDLE_ENFORCE_EQ
(
"If Attr(soft_label) == false, the axis dimension of "
labels_dims
[
axis
],
1UL
,
"Input(Label) should be 1."
);
platform
::
errors
::
InvalidArgument
(
"If Attr(soft_label) == false, "
"the axis dimension of "
"Input(Label) should be 1."
));
}
}
}
}
...
...
paddle/fluid/operators/softmax_with_cross_entropy_op.h
浏览文件 @
05c9642d
...
@@ -31,8 +31,9 @@ template <typename T>
...
@@ -31,8 +31,9 @@ template <typename T>
class
SoftmaxWithCrossEntropyKernel
:
public
framework
::
OpKernel
<
T
>
{
class
SoftmaxWithCrossEntropyKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
PADDLE_ENFORCE
(
platform
::
is_cpu_place
(
context
.
GetPlace
()),
PADDLE_ENFORCE_EQ
(
"This kernel only runs on CPU."
);
platform
::
is_cpu_place
(
context
.
GetPlace
()),
true
,
platform
::
errors
::
Unimplemented
(
"This kernel only runs on CPU."
));
const
Tensor
*
logits
=
context
.
Input
<
Tensor
>
(
"Logits"
);
const
Tensor
*
logits
=
context
.
Input
<
Tensor
>
(
"Logits"
);
const
Tensor
*
labels
=
context
.
Input
<
Tensor
>
(
"Label"
);
const
Tensor
*
labels
=
context
.
Input
<
Tensor
>
(
"Label"
);
Tensor
*
softmax
=
context
.
Output
<
Tensor
>
(
"Softmax"
);
Tensor
*
softmax
=
context
.
Output
<
Tensor
>
(
"Softmax"
);
...
...
paddle/fluid/operators/spp_op.cc
浏览文件 @
05c9642d
...
@@ -62,15 +62,17 @@ class SppOp : public framework::OperatorWithKernel {
...
@@ -62,15 +62,17 @@ class SppOp : public framework::OperatorWithKernel {
public:
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
"Input(X) of SppOp"
platform
::
errors
::
InvalidArgument
(
"should not be null."
);
"Input(X) of SppOp should not be null."
));
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
"Output(Out) of SppOp should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Output(Out) of SppOp should not be null."
));
auto
in_x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
in_x_dims
=
ctx
->
GetInputDim
(
"X"
);
int
pyramid_height
=
ctx
->
Attrs
().
Get
<
int
>
(
"pyramid_height"
);
int
pyramid_height
=
ctx
->
Attrs
().
Get
<
int
>
(
"pyramid_height"
);
PADDLE_ENFORCE
(
in_x_dims
.
size
()
==
4
,
PADDLE_ENFORCE_EQ
(
in_x_dims
.
size
(),
4
,
"Spping intput must be of 4-dimensional."
);
platform
::
errors
::
InvalidArgument
(
"Spping intput must be of 4-dimensional."
));
int
outlen
=
((
std
::
pow
(
4
,
pyramid_height
)
-
1
)
/
(
4
-
1
))
*
in_x_dims
[
1
];
int
outlen
=
((
std
::
pow
(
4
,
pyramid_height
)
-
1
)
/
(
4
-
1
))
*
in_x_dims
[
1
];
std
::
vector
<
int64_t
>
output_shape
({
in_x_dims
[
0
],
outlen
});
std
::
vector
<
int64_t
>
output_shape
({
in_x_dims
[
0
],
outlen
});
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
output_shape
));
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
output_shape
));
...
@@ -81,9 +83,12 @@ class SppOpGrad : public framework::OperatorWithKernel {
...
@@ -81,9 +83,12 @@ class SppOpGrad : public framework::OperatorWithKernel {
public:
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE
(
ctx
->
HasInput
(
"X"
),
"Input(X) must not be null."
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
ctx
->
HasInput
(
"X"
),
true
,
"Input(X@GRAD) should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Input(X) must not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
framework
::
GradVarName
(
"X"
)),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(X@GRAD) should not be null."
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
ctx
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
ctx
->
GetInputDim
(
"X"
));
}
}
};
};
...
...
paddle/fluid/operators/unsqueeze_op.cc
浏览文件 @
05c9642d
...
@@ -27,16 +27,22 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
...
@@ -27,16 +27,22 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
ctx
)
const
override
{
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasInput
(
"X"
),
true
,
"Input(X) of Unsqueeze operator should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Input(X) of "
"Unsqueeze operator should not be null."
));
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"Out"
),
true
,
"Output(Out) of Unsqueeze operator should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Output(Out) of "
"Unsqueeze operator should not be null."
));
const
auto
&
axes
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
int
>>
(
"axes"
);
const
auto
&
axes
=
ctx
->
Attrs
().
Get
<
std
::
vector
<
int
>>
(
"axes"
);
const
auto
&
x_dims
=
ctx
->
GetInputDim
(
"X"
);
const
auto
&
x_dims
=
ctx
->
GetInputDim
(
"X"
);
// Validity Check: input tensor dims (<6).
// Validity Check: input tensor dims (<6).
PADDLE_ENFORCE_LE
(
x_dims
.
size
(),
6
,
PADDLE_ENFORCE_LE
(
x_dims
.
size
(),
6
,
"Invalid dimensions, the rank of Input(X) "
platform
::
errors
::
InvalidArgument
(
"should be in the range of [1, 6] (Eigen limit)"
);
"Invalid "
"dimensions, the rank of Input(X) "
"should be in the range of [1, 6] (Eigen limit)"
));
if
(
!
axes
.
empty
())
{
if
(
!
axes
.
empty
())
{
auto
out_dims
=
GetOutputShape
(
axes
,
x_dims
);
auto
out_dims
=
GetOutputShape
(
axes
,
x_dims
);
ctx
->
SetOutputDim
(
"Out"
,
out_dims
);
ctx
->
SetOutputDim
(
"Out"
,
out_dims
);
...
@@ -49,24 +55,29 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
...
@@ -49,24 +55,29 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
auto
AxesTensorList
=
ctx
->
Inputs
(
"AxesTensorList"
);
auto
AxesTensorList
=
ctx
->
Inputs
(
"AxesTensorList"
);
int
output_size
=
x_dims
.
size
()
+
static_cast
<
int
>
(
AxesTensorList
.
size
());
int
output_size
=
x_dims
.
size
()
+
static_cast
<
int
>
(
AxesTensorList
.
size
());
PADDLE_ENFORCE_LE
(
output_size
,
6
,
PADDLE_ENFORCE_LE
(
output_size
,
6
,
"The output tensor's rank should be less than 6."
);
platform
::
errors
::
InvalidArgument
(
"The output tensor's rank should be less than 6."
));
std
::
vector
<
int
>
vec_out_dims
(
output_size
,
-
1
);
std
::
vector
<
int
>
vec_out_dims
(
output_size
,
-
1
);
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
vec_out_dims
));
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
vec_out_dims
));
}
else
if
(
ctx
->
HasInput
(
"AxesTensor"
))
{
}
else
if
(
ctx
->
HasInput
(
"AxesTensor"
))
{
auto
axes_dims
=
ctx
->
GetInputDim
(
"AxesTensor"
);
auto
axes_dims
=
ctx
->
GetInputDim
(
"AxesTensor"
);
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
axes_dims
.
size
(),
1
,
axes_dims
.
size
(),
1
,
platform
::
errors
::
InvalidArgument
(
"Input(AxesTensor)'s dimension of Op(unsqueeze) must be 1. "
"Input(AxesTensor)'s dimension of "
"But received AxesTensor's shape = [%s], "
"Op(unsqueeze) must be 1. "
"AxesTensor's dimension = %d."
,
"But received AxesTensor's shape = [%s], "
axes_dims
,
axes_dims
.
size
());
"AxesTensor's dimension = %d."
,
PADDLE_ENFORCE_GE
(
axes_dims
[
0
],
0
,
axes_dims
,
axes_dims
.
size
()));
"Input(AxesTensor)'s shape must be known. But received "
PADDLE_ENFORCE_GE
(
"AxesTensor's shape = [%s]"
,
axes_dims
[
0
],
0
,
axes_dims
);
platform
::
errors
::
InvalidArgument
(
"Input(AxesTensor)'s shape must be known. But received "
"AxesTensor's shape = [%s]"
,
axes_dims
));
int
output_size
=
x_dims
.
size
()
+
static_cast
<
int
>
(
axes_dims
[
0
]);
int
output_size
=
x_dims
.
size
()
+
static_cast
<
int
>
(
axes_dims
[
0
]);
PADDLE_ENFORCE_LE
(
output_size
,
6
,
PADDLE_ENFORCE_LE
(
output_size
,
6
,
"The output tensor's rank should be less than 6."
);
platform
::
errors
::
InvalidArgument
(
"The output tensor's rank should be less than 6."
));
std
::
vector
<
int
>
vec_out_dims
(
output_size
,
-
1
);
std
::
vector
<
int
>
vec_out_dims
(
output_size
,
-
1
);
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
vec_out_dims
));
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
vec_out_dims
));
}
}
...
@@ -80,13 +91,19 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
...
@@ -80,13 +91,19 @@ class UnsqueezeOp : public framework::OperatorWithKernel {
// Validity Check: rank range.
// Validity Check: rank range.
PADDLE_ENFORCE_LE
(
output_size
,
6
,
PADDLE_ENFORCE_LE
(
output_size
,
6
,
"The output tensor's rank should be less than 6."
);
platform
::
errors
::
InvalidArgument
(
"The output tensor's rank should be less than 6."
));
for
(
int
axis
:
unsqz_dims
)
{
for
(
int
axis
:
unsqz_dims
)
{
int
cur
=
axis
<
0
?
axis
+
cur_output_size
+
1
:
axis
;
int
cur
=
axis
<
0
?
axis
+
cur_output_size
+
1
:
axis
;
// Vaildity Check: the axis bound
// Vaildity Check: the axis bound
PADDLE_ENFORCE_GE
(
cur
,
0
);
PADDLE_ENFORCE_GE
(
cur
,
0
,
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_LE
(
cur
,
cur_output_size
);
"The insert dimension value should "
"not be less than 0"
));
PADDLE_ENFORCE_LE
(
cur
,
cur_output_size
,
platform
::
errors
::
InvalidArgument
(
"The insert dimension value shoud not be larger "
"than the dimension size of input tensor"
));
// Move old axis, and insert new axis
// Move old axis, and insert new axis
for
(
int
i
=
cur_output_size
;
i
>=
cur
;
--
i
)
{
for
(
int
i
=
cur_output_size
;
i
>=
cur
;
--
i
)
{
if
(
output_shape
[
i
]
==
1
)
{
if
(
output_shape
[
i
]
==
1
)
{
...
@@ -151,13 +168,17 @@ class UnsqueezeOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -151,13 +168,17 @@ class UnsqueezeOpMaker : public framework::OpProtoAndCheckerMaker {
.
AddCustomChecker
([](
const
std
::
vector
<
int
>
&
axes
)
{
.
AddCustomChecker
([](
const
std
::
vector
<
int
>
&
axes
)
{
// Validity Check: axes dims (<6).
// Validity Check: axes dims (<6).
PADDLE_ENFORCE_LT
(
static_cast
<
int
>
(
axes
.
size
()),
6
,
PADDLE_ENFORCE_LT
(
static_cast
<
int
>
(
axes
.
size
()),
6
,
"Invalid dimensions, dynamic dimensions should be "
platform
::
errors
::
InvalidArgument
(
"within [1, 6] dimensions (Eigen limit)."
);
"Invalid "
"dimensions, dynamic dimensions should be "
"within [1, 6] dimensions (Eigen limit)."
));
// Validity Check: the range of unsqueeze axis.
// Validity Check: the range of unsqueeze axis.
for
(
int
axis
:
axes
)
{
for
(
int
axis
:
axes
)
{
PADDLE_ENFORCE_LT
(
axis
,
6
,
PADDLE_ENFORCE_LT
(
axis
,
6
,
"Invalid dimensions, input axis should be"
platform
::
errors
::
InvalidArgument
(
" within [1, 6] dimensions (Eigen limit)."
);
"Invalid "
"dimensions, input axis should be"
"within [1, 6] dimensions (Eigen limit)."
));
}
}
});
});
AddComment
(
R"DOC(
AddComment
(
R"DOC(
...
@@ -219,7 +240,8 @@ class Unsqueeze2Op : public UnsqueezeOp {
...
@@ -219,7 +240,8 @@ class Unsqueeze2Op : public UnsqueezeOp {
PADDLE_ENFORCE_EQ
(
PADDLE_ENFORCE_EQ
(
ctx
->
HasOutput
(
"XShape"
),
true
,
ctx
->
HasOutput
(
"XShape"
),
true
,
"Output(XShape) of Unsqueeze operator should not be null."
);
platform
::
errors
::
InvalidArgument
(
"Output(XShape) of Unsqueeze "
"operator should not be null."
));
std
::
vector
<
int64_t
>
xshape_dims
(
x_dims
.
size
()
+
1
);
std
::
vector
<
int64_t
>
xshape_dims
(
x_dims
.
size
()
+
1
);
xshape_dims
[
0
]
=
0
;
xshape_dims
[
0
]
=
0
;
for
(
int
i
=
0
;
i
<
x_dims
.
size
();
++
i
)
{
for
(
int
i
=
0
;
i
<
x_dims
.
size
();
++
i
)
{
...
@@ -259,10 +281,12 @@ class Unsqueeze2GradOp : public framework::OperatorWithKernel {
...
@@ -259,10 +281,12 @@ class Unsqueeze2GradOp : public framework::OperatorWithKernel {
public:
public:
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
using
framework
::
OperatorWithKernel
::
OperatorWithKernel
;
void
InferShape
(
framework
::
InferShapeContext
*
context
)
const
override
{
void
InferShape
(
framework
::
InferShapeContext
*
context
)
const
override
{
PADDLE_ENFORCE_EQ
(
context
->
HasInput
(
"XShape"
),
true
,
PADDLE_ENFORCE_EQ
(
"Input(XShape) shouldn't be null."
);
context
->
HasInput
(
"XShape"
),
true
,
platform
::
errors
::
InvalidArgument
(
"Input(XShape) shouldn't be null."
));
PADDLE_ENFORCE_EQ
(
context
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
PADDLE_ENFORCE_EQ
(
context
->
HasInput
(
framework
::
GradVarName
(
"Out"
)),
true
,
"Input(Out@GRAD) shouldn't be null."
);
platform
::
errors
::
InvalidArgument
(
"Input(Out@GRAD) shouldn't be null."
));
auto
xshape_dims
=
context
->
GetInputDim
(
"XShape"
);
auto
xshape_dims
=
context
->
GetInputDim
(
"XShape"
);
auto
x_dims
=
framework
::
slice_ddim
(
xshape_dims
,
1
,
xshape_dims
.
size
());
auto
x_dims
=
framework
::
slice_ddim
(
xshape_dims
,
1
,
xshape_dims
.
size
());
context
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
x_dims
);
context
->
SetOutputDim
(
framework
::
GradVarName
(
"X"
),
x_dims
);
...
...
paddle/fluid/operators/unsqueeze_op.h
浏览文件 @
05c9642d
...
@@ -66,13 +66,20 @@ class UnsqueezeKernel : public framework::OpKernel<T> {
...
@@ -66,13 +66,20 @@ class UnsqueezeKernel : public framework::OpKernel<T> {
// Validity Check: rank range.
// Validity Check: rank range.
PADDLE_ENFORCE_LE
(
output_size
,
6
,
PADDLE_ENFORCE_LE
(
output_size
,
6
,
"The output tensor's rank should be less than 6."
);
platform
::
errors
::
InvalidArgument
(
"The output "
"tensor's rank should be less than 6."
));
for
(
int
axis
:
unsqz_dims
)
{
for
(
int
axis
:
unsqz_dims
)
{
int
cur
=
axis
<
0
?
axis
+
cur_output_size
+
1
:
axis
;
int
cur
=
axis
<
0
?
axis
+
cur_output_size
+
1
:
axis
;
// Vaildity Check: the axis bound
// Vaildity Check: the axis bound
PADDLE_ENFORCE_GE
(
cur
,
0
);
PADDLE_ENFORCE_GE
(
cur
,
0
,
platform
::
errors
::
InvalidArgument
(
PADDLE_ENFORCE_LE
(
cur
,
cur_output_size
);
"The insert dimension value should "
"not be less than 0"
));
PADDLE_ENFORCE_LE
(
cur
,
cur_output_size
,
platform
::
errors
::
InvalidArgument
(
"The insert dimension value shoule not be larger "
"than the dimension size of input tensor"
));
// Move old axis, and insert new axis
// Move old axis, and insert new axis
for
(
int
i
=
cur_output_size
;
i
>=
cur
;
--
i
)
{
for
(
int
i
=
cur_output_size
;
i
>=
cur
;
--
i
)
{
if
(
output_shape
[
i
]
==
1
)
{
if
(
output_shape
[
i
]
==
1
)
{
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录