Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
8de04be7
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
8de04be7
编写于
10月 18, 2017
作者:
W
wanghaoshuang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Fix unitest
上级
555ab3f5
变更
4
隐藏空白更改
内联
并排
Showing
4 changed file
with
64 addition
and
81 deletion
+64
-81
paddle/framework/lod_tensor.cc
paddle/framework/lod_tensor.cc
+29
-0
paddle/framework/lod_tensor.h
paddle/framework/lod_tensor.h
+7
-0
paddle/operators/seq_expand_op.h
paddle/operators/seq_expand_op.h
+21
-58
python/paddle/v2/framework/tests/test_seq_expand.py
python/paddle/v2/framework/tests/test_seq_expand.py
+7
-23
未找到文件。
paddle/framework/lod_tensor.cc
浏览文件 @
8de04be7
...
...
@@ -103,5 +103,34 @@ void LoDTensor::ShrinkInLevel(size_t level, size_t elem_begin,
lod_
=
new_lod
;
}
Vector
<
size_t
>
repeat_lod
(
Vector
<
size_t
>
data
,
Vector
<
size_t
>
starts
,
Vector
<
size_t
>
times
,
bool
is_first
)
{
Vector
<
size_t
>
result
;
result
.
push_back
(
data
[
0
]);
size_t
p
=
0
,
start
=
0
,
end
=
0
;
if
(
is_first
==
true
)
{
for
(
size_t
i
=
0
;
i
<
times
.
size
();
++
i
)
{
result
.
push_back
(
data
.
back
()
+
times
[
i
]
*
(
data
[
i
+
1
]
-
data
[
i
]));
}
}
else
{
for
(
size_t
i
=
0
;
i
<
times
.
size
();
++
i
)
{
while
(
starts
[
i
]
!=
data
[
p
]
&&
p
<
data
.
size
())
{
++
p
;
}
start
=
p
;
while
(
starts
[
i
+
1
]
!=
data
[
p
]
&&
p
<
data
.
size
())
{
++
p
;
}
end
=
p
+
1
;
for
(
size_t
j
=
0
;
j
<
times
[
i
];
++
j
)
{
for
(
size_t
index
=
start
;
index
<
end
-
1
;
++
index
)
{
result
.
push_back
(
result
.
back
()
+
data
[
index
+
1
]
-
data
[
index
]);
}
}
}
}
return
result
;
}
}
// namespace framework
}
// namespace paddle
paddle/framework/lod_tensor.h
浏览文件 @
8de04be7
...
...
@@ -15,6 +15,9 @@
#pragma once
#include <memory>
#include "paddle/memory/memcpy.h"
#include "paddle/platform/device_context.h"
#include "paddle/platform/place.h"
#ifdef PADDLE_WITH_CUDA
#include <thrust/device_vector.h>
#include <thrust/host_vector.h>
...
...
@@ -122,5 +125,9 @@ class LoDTensor : public Tensor {
private:
LoD
lod_
;
};
Vector
<
size_t
>
repeat_lod
(
Vector
<
size_t
>
data
,
Vector
<
size_t
>
starts
,
Vector
<
size_t
>
times
,
bool
is_first
);
}
// namespace framework
}
// namespace paddle
paddle/operators/seq_expand_op.h
浏览文件 @
8de04be7
...
...
@@ -22,54 +22,6 @@ namespace operators {
using
LoDTensor
=
framework
::
LoDTensor
;
template
<
typename
T
>
using
vector
=
framework
::
Vector
<
T
>
;
vector
<
size_t
>
repeat_lod
(
vector
<
size_t
>
data
,
vector
<
size_t
>
starts
,
vector
<
size_t
>
times
,
bool
is_first
)
{
vector
<
size_t
>
result
;
result
.
push_back
(
data
[
0
]);
size_t
p
=
0
,
start
=
0
,
end
=
0
;
if
(
is_first
==
true
)
{
for
(
size_t
i
=
0
;
i
<
times
.
size
();
++
i
)
{
result
.
push_back
(
data
.
back
()
+
times
[
i
]
*
(
data
[
i
+
1
]
-
data
[
i
]));
}
}
else
{
for
(
size_t
i
=
0
;
i
<
times
.
size
();
++
i
)
{
while
(
starts
[
i
]
!=
data
[
p
]
&&
p
<
data
.
size
())
{
++
p
;
}
start
=
p
;
while
(
starts
[
i
+
1
]
!=
data
[
p
]
&&
p
<
data
.
size
())
{
++
p
;
}
end
=
p
+
1
;
for
(
size_t
j
=
0
;
j
<
times
[
i
];
++
j
)
{
for
(
size_t
index
=
start
;
index
<
end
-
1
;
++
index
)
{
result
.
push_back
(
result
.
back
()
+
data
[
index
+
1
]
-
data
[
index
]);
}
}
}
}
return
result
;
}
template
<
typename
Place
,
typename
T
>
void
repeat_data
(
const
T
*
src
,
T
*
dst
,
size_t
size
,
vector
<
size_t
>
starts
,
vector
<
size_t
>
times
,
Place
place
)
{
const
T
*
src_p
=
src
;
T
*
dst_p
=
dst
;
size_t
count
=
0
;
for
(
size_t
i
=
0
;
i
<
times
.
size
();
++
i
)
{
count
=
size
*
(
starts
[
i
+
1
]
-
starts
[
i
]);
for
(
size_t
j
=
0
;
j
<
times
[
i
];
++
j
)
{
memory
::
Copy
(
place
,
dst_p
,
place
,
src_p
,
sizeof
(
T
)
*
count
);
dst_p
+=
count
;
}
src_p
+=
count
;
}
}
template
<
typename
Place
,
typename
T
>
class
SeqExpandKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
...
...
@@ -81,7 +33,7 @@ class SeqExpandKernel : public framework::OpKernel<T> {
auto
x_lod
=
x
->
lod
();
if
(
x_lod
.
size
()
==
0
)
{
v
ector
<
size_t
>
level
;
framework
::
V
ector
<
size_t
>
level
;
for
(
int
i
=
0
;
i
<
x
->
dims
()[
0
]
+
1
;
++
i
)
{
level
.
push_back
(
i
);
}
...
...
@@ -91,7 +43,7 @@ class SeqExpandKernel : public framework::OpKernel<T> {
}
size_t
repeat
=
static_cast
<
size_t
>
(
context
.
Attr
<
int
>
(
"repeat"
));
v
ector
<
size_t
>
repeats
;
framework
::
V
ector
<
size_t
>
repeats
;
if
(
repeat
!=
0
)
{
for
(
int
i
=
0
;
i
<
x_lod
[
0
].
size
()
-
1
;
++
i
)
{
repeats
.
push_back
(
repeat
);
...
...
@@ -107,21 +59,32 @@ class SeqExpandKernel : public framework::OpKernel<T> {
repeats
.
push_back
((
y_lod
[
0
][
i
+
1
]
-
y_lod
[
0
][
i
])
/
(
x_lod
[
0
][
i
+
1
]
-
x_lod
[
0
][
i
]));
}
out
->
Resize
(
x_dims
);
out
->
Resize
(
y
->
dims
()
);
}
framework
::
LoD
out_lod
;
auto
level0
=
repeat_lod
(
x_lod
[
0
],
x_lod
[
0
],
repeats
,
true
);
auto
level0
=
framework
::
repeat_lod
(
x_lod
[
0
],
x_lod
[
0
],
repeats
,
true
);
out_lod
.
push_back
(
level0
);
for
(
int
i
=
1
;
i
<
x_lod
.
size
();
++
i
)
{
out_lod
.
push_back
(
repeat_lod
(
x_lod
[
i
],
x_lod
[
0
],
repeats
,
false
));
out_lod
.
push_back
(
framework
::
repeat_lod
(
x_lod
[
i
],
x_lod
[
0
],
repeats
,
false
));
}
size_t
element_len
=
framework
::
product
(
x_dims
)
/
x_dims
[
0
];
T
*
out_data
=
out
->
mutable_data
<
T
>
(
context
.
GetPlace
());
// copy data
Place
place
=
boost
::
get
<
Place
>
(
context
.
GetPlace
());
repeat_data
<
Place
,
T
>
(
x_data
,
out_data
,
element_len
,
x_lod
[
0
],
repeats
,
place
);
size_t
count
=
0
;
for
(
size_t
i
=
0
;
i
<
repeats
.
size
();
++
i
)
{
count
=
element_len
*
(
x_lod
[
0
][
i
+
1
]
-
x_lod
[
0
][
i
]);
for
(
size_t
j
=
0
;
j
<
repeats
[
i
];
++
j
)
{
memory
::
Copy
(
place
,
out_data
,
place
,
x_data
,
sizeof
(
T
)
*
count
);
out_data
+=
count
;
}
x_data
+=
count
;
}
out
->
set_lod
(
out_lod
);
}
};
...
...
@@ -130,9 +93,9 @@ template <typename Place, typename T>
class
SeqExpandGradKernel
:
public
framework
::
OpKernel
<
T
>
{
public:
void
Compute
(
const
framework
::
ExecutionContext
&
context
)
const
override
{
//
auto* d_out = context.Input<Tensor>(framework::GradVarName("Out"));
//
auto* d_x = context.Output<Tensor>(framework::GradVarName("X"));
//
d_x->mutable_data<T>(context.GetPlace());
auto
*
d_out
=
context
.
Input
<
Tensor
>
(
framework
::
GradVarName
(
"Out"
));
auto
*
d_x
=
context
.
Output
<
Tensor
>
(
framework
::
GradVarName
(
"X"
));
d_x
->
mutable_data
<
T
>
(
context
.
GetPlace
());
}
};
...
...
python/paddle/v2/framework/tests/test_seq_expand.py
浏览文件 @
8de04be7
...
...
@@ -29,17 +29,13 @@ def repeat_array(array, starts, times):
class
TestSeqExpand
(
OpTest
):
def
set_data
(
self
):
self
.
op_type
=
'seq_expand'
x
=
np
.
random
.
uniform
(
0.1
,
1
,
[
3
,
2
,
2
]).
astype
(
'float32'
)
y
=
np
.
zeros
((
6
,
2
,
2
)).
astype
(
'float32'
)
y_lod
=
[[
0
,
2
,
3
,
6
]]
self
.
inputs
=
{
'X'
:
(
x
,
None
),
'Y'
:
(
y
,
y_lod
)}
x_data
=
np
.
random
.
uniform
(
0.1
,
1
,
[
4
,
1
]).
astype
(
'float32'
)
self
.
inputs
=
{
'X'
:
x_data
}
self
.
repeat
=
2
def
compute
(
self
):
x_data
,
x_lod
=
self
.
inputs
[
'X'
]
print
"x_data: %s"
%
x_data
print
"x_lod: %s"
%
x_lod
x
=
self
.
inputs
[
'X'
]
x_data
,
x_lod
=
x
if
type
(
x
)
==
tuple
else
(
x
,
None
)
if
not
x_lod
:
x_lod
=
[[
i
for
i
in
range
(
1
+
x_data
.
shape
[
0
])]]
else
:
...
...
@@ -47,28 +43,16 @@ class TestSeqExpand(OpTest):
if
self
.
repeat
:
self
.
attrs
=
{
'repeat'
:
self
.
repeat
}
repeats
=
(
len
(
x_lod
[
0
])
-
1
)
*
[
self
.
repeat
]
# get out shape
# out_shape = np.copy(x_data.shape)
# out_shape[0] = out_shape[0] * self.repeat
else
:
y_data
,
y_lod
=
self
.
inputs
[
'Y'
]
print
"y_lod: %s"
%
y_lod
#print "y_lod: %s" % y_lod
# get repeats
repeats
=
[((
y_lod
[
0
][
i
+
1
]
-
y_lod
[
0
][
i
])
/
(
x_lod
[
0
][
i
+
1
]
-
x_lod
[
0
][
i
]))
for
i
in
range
(
len
(
y_lod
[
0
])
-
1
)]
# get out shape
# out_shape = y_data.shape
# get out lod
out_lod
=
[
repeat
(
x_lod
[
0
],
x_lod
[
0
],
repeats
,
True
)]
+
[
repeat
(
lod
,
x_lod
[
0
],
repeats
,
False
)
for
lod
in
x_lod
[
1
:]
]
# copy data
out
=
repeat_array
(
x_data
.
tolist
(),
x_lod
[
0
],
repeats
)
self
.
outputs
=
{
'Out'
:
(
out
,
out_lod
)}
print
"outputs: %s"
%
self
.
outputs
self
.
outputs
=
{
'Out'
:
out
}
def
setUp
(
self
):
self
.
op_type
=
'seq_expand'
...
...
@@ -94,7 +78,7 @@ class TestSeqExpandCase1(TestSeqExpand):
class
TestSeqExpandCase2
(
TestSeqExpand
):
def
set_data
(
self
):
x_data
=
np
.
random
.
uniform
(
0.1
,
1
,
[
4
,
1
]).
astype
(
'float32'
)
self
.
inputs
=
{
'X'
:
(
x_data
,
None
)
}
self
.
inputs
=
{
'X'
:
x_data
}
self
.
repeat
=
2
...
...
@@ -103,7 +87,7 @@ class TestSeqExpandCase3(TestSeqExpand):
x_data
=
np
.
random
.
uniform
(
0.1
,
1
,
[
3
,
1
]).
astype
(
'float32'
)
y_data
=
np
.
random
.
uniform
(
0.1
,
1
,
[
8
,
1
]).
astype
(
'float32'
)
y_lod
=
[[
0
,
1
,
4
,
8
]]
self
.
inputs
=
{
'X'
:
(
x_data
,
None
)
,
'Y'
:
(
y_data
,
y_lod
)}
self
.
inputs
=
{
'X'
:
x_data
,
'Y'
:
(
y_data
,
y_lod
)}
self
.
repeat
=
None
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录