Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
9cc1937d
P
Paddle
项目概览
PaddlePaddle
/
Paddle
大约 1 年 前同步成功
通知
2299
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
9cc1937d
编写于
9月 18, 2018
作者:
Y
Yibing Liu
提交者:
GitHub
9月 18, 2018
浏览文件
操作
浏览文件
下载
差异文件
Merge pull request #13400 from kuke/fix_seq_pad
Get sequence length in sequence_pad op & fix sequence_mask op
上级
020d13c1
1c87558c
变更
6
隐藏空白更改
内联
并排
Showing
6 changed file
with
69 addition
and
13 deletion
+69
-13
paddle/fluid/operators/sequence_mask_op.cc
paddle/fluid/operators/sequence_mask_op.cc
+5
-1
paddle/fluid/operators/sequence_mask_op.cu
paddle/fluid/operators/sequence_mask_op.cu
+5
-1
paddle/fluid/operators/sequence_pad_op.cc
paddle/fluid/operators/sequence_pad_op.cc
+32
-6
paddle/fluid/operators/sequence_pad_op.h
paddle/fluid/operators/sequence_pad_op.h
+10
-0
python/paddle/fluid/layers/nn.py
python/paddle/fluid/layers/nn.py
+11
-4
python/paddle/fluid/tests/unittests/test_sequence_pad_op.py
python/paddle/fluid/tests/unittests/test_sequence_pad_op.py
+6
-1
未找到文件。
paddle/fluid/operators/sequence_mask_op.cc
浏览文件 @
9cc1937d
...
@@ -23,4 +23,8 @@ REGISTER_OP_CPU_KERNEL(
...
@@ -23,4 +23,8 @@ REGISTER_OP_CPU_KERNEL(
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int
>
,
int
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
int64_t
>
);
int64_t
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
float
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CPUDeviceContext
,
double
>
);
paddle/fluid/operators/sequence_mask_op.cu
浏览文件 @
9cc1937d
...
@@ -19,4 +19,8 @@ REGISTER_OP_CUDA_KERNEL(
...
@@ -19,4 +19,8 @@ REGISTER_OP_CUDA_KERNEL(
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int
>
,
int
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
int64_t
>
);
int64_t
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
float
>
,
paddle
::
operators
::
SequenceMaskKernel
<
paddle
::
platform
::
CUDADeviceContext
,
double
>
);
paddle/fluid/operators/sequence_pad_op.cc
浏览文件 @
9cc1937d
...
@@ -29,10 +29,12 @@ class SequencePadOp : public framework::OperatorWithKernel {
...
@@ -29,10 +29,12 @@ class SequencePadOp : public framework::OperatorWithKernel {
"Input(PadValue) of SequencePadOp should not be null."
);
"Input(PadValue) of SequencePadOp should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Out"
),
"Output(Out) of SequencePadOp should not be null."
);
"Output(Out) of SequencePadOp should not be null."
);
PADDLE_ENFORCE
(
ctx
->
HasOutput
(
"Length"
),
"Output(Length) of SequencePadOp should not be null."
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
auto
x_dims
=
ctx
->
GetInputDim
(
"X"
);
PADDLE_ENFORCE_GE
(
x_dims
.
size
(),
2
,
PADDLE_ENFORCE_GE
(
x_dims
.
size
(),
2
,
"The rank of Input(
x
) can't be less than 2."
);
"The rank of Input(
X
) can't be less than 2."
);
auto
time_step_dims
=
framework
::
slice_ddim
(
x_dims
,
1
,
x_dims
.
size
());
auto
time_step_dims
=
framework
::
slice_ddim
(
x_dims
,
1
,
x_dims
.
size
());
auto
pad_value_dims
=
ctx
->
GetInputDim
(
"PadValue"
);
auto
pad_value_dims
=
ctx
->
GetInputDim
(
"PadValue"
);
PADDLE_ENFORCE
(
pad_value_dims
==
framework
::
make_ddim
({
1
})
||
PADDLE_ENFORCE
(
pad_value_dims
==
framework
::
make_ddim
({
1
})
||
...
@@ -41,8 +43,8 @@ class SequencePadOp : public framework::OperatorWithKernel {
...
@@ -41,8 +43,8 @@ class SequencePadOp : public framework::OperatorWithKernel {
"shape equals to time steps in sequences"
);
"shape equals to time steps in sequences"
);
int
out_dim_0
=
-
1
;
int
out_dim_0
=
-
1
;
int
out_dim_1
=
-
1
;
int
padded_length
=
ctx
->
Attrs
().
Get
<
int
>
(
"padded_length"
);
if
(
ctx
->
IsRuntime
())
{
if
(
ctx
->
IsRuntime
())
{
// run time
// run time
framework
::
Variable
*
x_var
=
framework
::
Variable
*
x_var
=
...
@@ -58,7 +60,6 @@ class SequencePadOp : public framework::OperatorWithKernel {
...
@@ -58,7 +60,6 @@ class SequencePadOp : public framework::OperatorWithKernel {
int
seq_num
=
x_lod_0
.
size
()
-
1
;
int
seq_num
=
x_lod_0
.
size
()
-
1
;
int
max_seq_len
=
math
::
MaximumSequenceLength
(
x_lod_0
);
int
max_seq_len
=
math
::
MaximumSequenceLength
(
x_lod_0
);
int
padded_length
=
ctx
->
Attrs
().
Get
<
int
>
(
"padded_length"
);
if
(
padded_length
==
-
1
)
{
if
(
padded_length
==
-
1
)
{
padded_length
=
max_seq_len
;
padded_length
=
max_seq_len
;
}
}
...
@@ -66,19 +67,30 @@ class SequencePadOp : public framework::OperatorWithKernel {
...
@@ -66,19 +67,30 @@ class SequencePadOp : public framework::OperatorWithKernel {
"The Attr(padded_length) must be -1 or an int greater "
"The Attr(padded_length) must be -1 or an int greater "
"than the length of the longest original sequence."
);
"than the length of the longest original sequence."
);
out_dim_0
=
seq_num
;
out_dim_0
=
seq_num
;
out_dim_1
=
padded_length
;
}
else
{
}
else
{
// compile time
// compile time
if
(
padded_length
==
-
1
)
{
padded_length
=
1
;
}
framework
::
VarDesc
*
x_desc
=
framework
::
VarDesc
*
x_desc
=
boost
::
get
<
framework
::
VarDesc
*>
(
ctx
->
GetInputVarPtrs
(
"X"
)[
0
]);
boost
::
get
<
framework
::
VarDesc
*>
(
ctx
->
GetInputVarPtrs
(
"X"
)[
0
]);
PADDLE_ENFORCE_GE
(
x_desc
->
GetLoDLevel
(),
1
);
PADDLE_ENFORCE_GE
(
x_desc
->
GetLoDLevel
(),
1
);
}
}
std
::
vector
<
int
>
out_dims_vec
{
out_dim_0
,
out_dim_1
};
std
::
vector
<
int
>
out_dims_vec
{
out_dim_0
,
padded_length
};
std
::
vector
<
int
>
len_dims_vec
{
out_dim_0
,
1
};
auto
time_step_dims_vec
=
framework
::
vectorize2int
(
time_step_dims
);
auto
time_step_dims_vec
=
framework
::
vectorize2int
(
time_step_dims
);
out_dims_vec
.
insert
(
out_dims_vec
.
end
(),
time_step_dims_vec
.
begin
(),
out_dims_vec
.
insert
(
out_dims_vec
.
end
(),
time_step_dims_vec
.
begin
(),
time_step_dims_vec
.
end
());
time_step_dims_vec
.
end
());
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
out_dims_vec
));
ctx
->
SetOutputDim
(
"Out"
,
framework
::
make_ddim
(
out_dims_vec
));
ctx
->
SetOutputDim
(
"Length"
,
framework
::
make_ddim
(
len_dims_vec
));
}
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
data_type
=
framework
::
GetDataTypeOfVar
(
ctx
.
InputVar
(
"X"
));
return
framework
::
OpKernelType
(
data_type
,
ctx
.
device_context
());
}
}
};
};
...
@@ -96,6 +108,10 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -96,6 +108,10 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
AddOutput
(
AddOutput
(
"Out"
,
"Out"
,
"(LoDTensor) The output vairable, which contains padded sequences."
);
"(LoDTensor) The output vairable, which contains padded sequences."
);
AddOutput
(
"Length"
,
"(LoDTensor) The output vairable, which contains the actual length of "
"sequences before padding."
);
AddAttr
<
int
>
(
AddAttr
<
int
>
(
"padded_length"
,
"padded_length"
,
"The length of padded sequences. It can be setted to -1 or "
"The length of padded sequences. It can be setted to -1 or "
...
@@ -125,6 +141,7 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -125,6 +141,7 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
then we get LoDTensor:
then we get LoDTensor:
Out.data = [[a, b, 0, 0],
Out.data = [[a, b, 0, 0],
[c, d, e, 0]]
[c, d, e, 0]]
Length.data = [[2], [3]]
Case 2:
Case 2:
...
@@ -138,7 +155,8 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -138,7 +155,8 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
then we get LoDTensor:
then we get LoDTensor:
Out.data = [[[a1, a2], [b1, b2], [0, 0]],
Out.data = [[[a1, a2], [b1, b2], [0, 0]],
[[c1, c2], [d1, d2], [e1, e2]]]
[[c1, c2], [d1, d2], [e1, e2]]]
Length.data = [[2], [3]]
Case 3:
Case 3:
Given a 1-level LoDTensor input(X):
Given a 1-level LoDTensor input(X):
...
@@ -151,6 +169,7 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
...
@@ -151,6 +169,7 @@ class SequencePadOpMaker : public framework::OpProtoAndCheckerMaker {
then we get LoDTensor:
then we get LoDTensor:
Out.data = [[[a1, a2], [b1, b2], [p1, p2]],
Out.data = [[[a1, a2], [b1, b2], [p1, p2]],
[[c1, c2], [d1, d2], [e1, e2]]]
[[c1, c2], [d1, d2], [e1, e2]]]
Length.data = [[2], [3]]
)DOC"
);
)DOC"
);
}
}
...
@@ -171,6 +190,13 @@ class SequencePadGradOp : public framework::OperatorWithKernel {
...
@@ -171,6 +190,13 @@ class SequencePadGradOp : public framework::OperatorWithKernel {
ctx
->
ShareLoD
(
"X"
,
/*->*/
framework
::
GradVarName
(
"X"
));
ctx
->
ShareLoD
(
"X"
,
/*->*/
framework
::
GradVarName
(
"X"
));
}
}
}
}
protected:
framework
::
OpKernelType
GetExpectedKernelType
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
auto
data_type
=
framework
::
GetDataTypeOfVar
(
ctx
.
InputVar
(
"X"
));
return
framework
::
OpKernelType
(
data_type
,
ctx
.
device_context
());
}
};
};
}
// namespace operators
}
// namespace operators
...
...
paddle/fluid/operators/sequence_pad_op.h
浏览文件 @
9cc1937d
...
@@ -32,6 +32,7 @@ class SequencePadOpKernel : public framework::OpKernel<T> {
...
@@ -32,6 +32,7 @@ class SequencePadOpKernel : public framework::OpKernel<T> {
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
const
auto
*
x
=
ctx
.
Input
<
LoDTensor
>
(
"X"
);
const
auto
*
x
=
ctx
.
Input
<
LoDTensor
>
(
"X"
);
auto
*
out
=
ctx
.
Output
<
LoDTensor
>
(
"Out"
);
auto
*
out
=
ctx
.
Output
<
LoDTensor
>
(
"Out"
);
auto
*
len_t
=
ctx
.
Output
<
LoDTensor
>
(
"Length"
);
out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
out
->
mutable_data
<
T
>
(
ctx
.
GetPlace
());
const
auto
*
pad_value
=
ctx
.
Input
<
LoDTensor
>
(
"PadValue"
);
const
auto
*
pad_value
=
ctx
.
Input
<
LoDTensor
>
(
"PadValue"
);
...
@@ -41,6 +42,15 @@ class SequencePadOpKernel : public framework::OpKernel<T> {
...
@@ -41,6 +42,15 @@ class SequencePadOpKernel : public framework::OpKernel<T> {
math
::
PaddingLoDTensorFunctor
<
DeviceContext
,
T
>
()(
math
::
PaddingLoDTensorFunctor
<
DeviceContext
,
T
>
()(
ctx
.
template
device_context
<
DeviceContext
>(),
*
x
,
out
,
*
pad_value
,
ctx
.
template
device_context
<
DeviceContext
>(),
*
x
,
out
,
*
pad_value
,
padded_length
,
0
,
false
,
math
::
kBatchLengthWidth
);
padded_length
,
0
,
false
,
math
::
kBatchLengthWidth
);
LoDTensor
seq_len
;
seq_len
.
Resize
(
len_t
->
dims
());
int64_t
*
len_data
=
seq_len
.
mutable_data
<
int64_t
>
(
platform
::
CPUPlace
());
for
(
size_t
i
=
1
;
i
<
x
->
lod
()[
0
].
size
();
++
i
)
{
len_data
[
i
-
1
]
=
x
->
lod
()[
0
][
i
]
-
x
->
lod
()[
0
][
i
-
1
];
}
framework
::
TensorCopy
(
seq_len
,
ctx
.
GetPlace
(),
ctx
.
template
device_context
<
DeviceContext
>(),
len_t
);
}
}
};
};
...
...
python/paddle/fluid/layers/nn.py
浏览文件 @
9cc1937d
...
@@ -2776,7 +2776,8 @@ def sequence_pad(x, pad_value, maxlen=None):
...
@@ -2776,7 +2776,8 @@ def sequence_pad(x, pad_value, maxlen=None):
longest original sequence."
longest original sequence."
Returns:
Returns:
Variable: The padded sequence batch. All sequences has the same length.
Variable: The padded sequence batch and the original lengths before
padding. All sequences has the same length.
Examples:
Examples:
.. code-block:: python
.. code-block:: python
...
@@ -2792,15 +2793,21 @@ def sequence_pad(x, pad_value, maxlen=None):
...
@@ -2792,15 +2793,21 @@ def sequence_pad(x, pad_value, maxlen=None):
helper
=
LayerHelper
(
'sequence_pad'
,
input
=
x
,
**
locals
())
helper
=
LayerHelper
(
'sequence_pad'
,
input
=
x
,
**
locals
())
dtype
=
helper
.
input_dtype
()
dtype
=
helper
.
input_dtype
()
out
=
helper
.
create_tmp_variable
(
dtype
)
out
=
helper
.
create_tmp_variable
(
dtype
)
length
=
helper
.
create_tmp_variable
(
dtype
)
pad_value
.
stop_gradient
=
True
length
.
stop_gradient
=
True
if
maxlen
is
None
:
if
maxlen
is
None
:
maxlen
=
-
1
maxlen
=
-
1
helper
.
append_op
(
helper
.
append_op
(
type
=
'sequence_pad'
,
type
=
'sequence_pad'
,
inputs
=
{
'X'
:
x
,
inputs
=
{
'X'
:
x
,
'PadValue'
:
pad_value
},
'PadValue'
:
pad_value
},
outputs
=
{
'Out'
:
out
},
outputs
=
{
'Out'
:
out
,
'Length'
:
length
},
attrs
=
{
'padded_length'
:
maxlen
})
attrs
=
{
'padded_length'
:
maxlen
})
return
out
return
out
,
length
def
beam_search
(
pre_ids
,
def
beam_search
(
pre_ids
,
...
@@ -6017,7 +6024,7 @@ def sequence_mask(x, maxlen=None, dtype='int64', name=None):
...
@@ -6017,7 +6024,7 @@ def sequence_mask(x, maxlen=None, dtype='int64', name=None):
inputs
=
{
'X'
:
[
x
]},
inputs
=
{
'X'
:
[
x
]},
outputs
=
{
'Y'
:
out
},
outputs
=
{
'Y'
:
out
},
attrs
=
{
attrs
=
{
'max
_
len'
:
maxlen
if
maxlen
is
not
None
else
-
1
,
'maxlen'
:
maxlen
if
maxlen
is
not
None
else
-
1
,
'out_dtype'
:
out
.
dtype
'out_dtype'
:
out
.
dtype
})
})
return
out
return
out
...
...
python/paddle/fluid/tests/unittests/test_sequence_pad_op.py
浏览文件 @
9cc1937d
...
@@ -62,7 +62,8 @@ class TestSequencePadOp(OpTest):
...
@@ -62,7 +62,8 @@ class TestSequencePadOp(OpTest):
start_idx
=
end_idx
start_idx
=
end_idx
out_data
=
np
.
array
(
padded_sequences
)
out_data
=
np
.
array
(
padded_sequences
)
self
.
outputs
=
{
'Out'
:
out_data
}
length
=
np
.
array
(
self
.
x_len_lod
[
0
]).
reshape
((
-
1
,
1
))
self
.
outputs
=
{
'Out'
:
out_data
,
'Length'
:
length
}
def
setUp
(
self
):
def
setUp
(
self
):
self
.
op_type
=
'sequence_pad'
self
.
op_type
=
'sequence_pad'
...
@@ -129,3 +130,7 @@ class TestSequencePadOp7(TestSequencePadOp):
...
@@ -129,3 +130,7 @@ class TestSequencePadOp7(TestSequencePadOp):
self
.
pad_value
=
[
1.0
]
self
.
pad_value
=
[
1.0
]
self
.
padded_length
=
7
self
.
padded_length
=
7
self
.
dtype
=
'float32'
self
.
dtype
=
'float32'
if
__name__
==
'__main__'
:
unittest
.
main
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录