Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
机器未来
Paddle
提交
8c23f7c4
P
Paddle
项目概览
机器未来
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
8c23f7c4
编写于
8月 03, 2018
作者:
T
tensor-tang
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
fix blas and use packed weight
上级
d9cc6b18
变更
2
隐藏空白更改
内联
并排
Showing
2 changed file
with
28 addition
and
8 deletion
+28
-8
paddle/fluid/operators/gru_op.h
paddle/fluid/operators/gru_op.h
+27
-7
paddle/fluid/operators/math/blas.h
paddle/fluid/operators/math/blas.h
+1
-1
未找到文件。
paddle/fluid/operators/gru_op.h
浏览文件 @
8c23f7c4
...
@@ -98,6 +98,23 @@ class GRUKernel : public framework::OpKernel<T> {
...
@@ -98,6 +98,23 @@ class GRUKernel : public framework::OpKernel<T> {
auto
active_gate
=
math
::
detail
::
GetActivationType
(
auto
active_gate
=
math
::
detail
::
GetActivationType
(
context
.
Attr
<
std
::
string
>
(
"gate_activation"
));
context
.
Attr
<
std
::
string
>
(
"gate_activation"
));
auto
blas
=
math
::
GetBlas
<
DeviceContext
,
T
>
(
dev_ctx
);
auto
blas
=
math
::
GetBlas
<
DeviceContext
,
T
>
(
dev_ctx
);
// TODO(TJ): make a class, make one pack
T
*
packed_gate
=
blas
.
GEMM_ALLOC
(
CblasBMatrix
,
1
/*height of C*/
,
frame_size
*
2
/*width of weight*/
,
frame_size
/*height of height*/
);
PADDLE_ENFORCE
(
packed_gate
);
blas
.
GEMM_PACK
(
CblasBMatrix
,
CblasNoTrans
,
1
/*cur bs?*/
,
frame_size
*
2
,
frame_size
,
T
(
1.0
),
gru_value
.
gate_weight
,
frame_size
*
2
,
packed_gate
);
T
*
packed_state
=
blas
.
GEMM_ALLOC
(
CblasBMatrix
,
1
/*height of C*/
,
frame_size
/*width of weight*/
,
frame_size
/*height of height*/
);
PADDLE_ENFORCE
(
packed_state
);
blas
.
GEMM_PACK
(
CblasBMatrix
,
CblasNoTrans
,
1
/*cur bs?*/
,
frame_size
,
frame_size
,
T
(
1.0
),
gru_value
.
state_weight
,
frame_size
,
packed_state
);
for
(
size_t
n
=
0
;
n
<
num_batch
;
n
++
)
{
for
(
size_t
n
=
0
;
n
<
num_batch
;
n
++
)
{
int
bstart
=
static_cast
<
int
>
(
batch_starts
[
n
]);
int
bstart
=
static_cast
<
int
>
(
batch_starts
[
n
]);
int
bend
=
static_cast
<
int
>
(
batch_starts
[
n
+
1
]);
int
bend
=
static_cast
<
int
>
(
batch_starts
[
n
+
1
]);
...
@@ -110,9 +127,10 @@ class GRUKernel : public framework::OpKernel<T> {
...
@@ -110,9 +127,10 @@ class GRUKernel : public framework::OpKernel<T> {
gru_value
.
gate_value
=
gate_t
.
data
<
T
>
();
gru_value
.
gate_value
=
gate_t
.
data
<
T
>
();
gru_value
.
reset_output_value
=
reset_hidden_prev_t
.
data
<
T
>
();
gru_value
.
reset_output_value
=
reset_hidden_prev_t
.
data
<
T
>
();
if
(
gru_value
.
prev_out_value
)
{
if
(
gru_value
.
prev_out_value
)
{
blas
.
GEMM
(
false
,
false
,
cur_batch_size
,
frame_size
*
2
,
frame_size
,
1
,
blas
.
GEMM_COMPUTE
(
CblasNoTrans
,
CblasPacked
,
cur_batch_size
,
gru_value
.
prev_out_value
,
frame_size
,
gru_value
.
gate_weight
,
frame_size
*
2
,
frame_size
,
gru_value
.
prev_out_value
,
frame_size
*
2
,
1
,
gru_value
.
gate_value
,
frame_size
*
3
);
frame_size
,
packed_gate
,
frame_size
*
2
,
T
(
1
),
gru_value
.
gate_value
,
frame_size
*
3
);
}
}
math
::
detail
::
forward_reset_output
(
math
::
detail
::
forward_reset_output
(
...
@@ -120,10 +138,10 @@ class GRUKernel : public framework::OpKernel<T> {
...
@@ -120,10 +138,10 @@ class GRUKernel : public framework::OpKernel<T> {
cur_batch_size
,
active_gate
);
cur_batch_size
,
active_gate
);
if
(
gru_value
.
prev_out_value
)
{
if
(
gru_value
.
prev_out_value
)
{
blas
.
GEMM
(
false
,
false
,
cur_batch_size
,
frame_size
,
frame_size
,
1
,
blas
.
GEMM
_COMPUTE
(
gru_value
.
reset_output_valu
e
,
frame_size
,
CblasNoTrans
,
CblasPacked
,
cur_batch_size
,
frame_siz
e
,
frame_size
,
gru_value
.
state_weight
,
frame_size
,
1
,
gru_value
.
reset_output_value
,
frame_size
,
packed_state
,
frame_size
,
gru_value
.
gate_value
+
frame_size
*
2
,
frame_size
*
3
);
T
(
1
),
gru_value
.
gate_value
+
frame_size
*
2
,
frame_size
*
3
);
}
}
math
::
detail
::
forward_final_output
(
math
::
detail
::
forward_final_output
(
...
@@ -132,6 +150,8 @@ class GRUKernel : public framework::OpKernel<T> {
...
@@ -132,6 +150,8 @@ class GRUKernel : public framework::OpKernel<T> {
gru_value
.
prev_out_value
=
gru_value
.
output_value
;
gru_value
.
prev_out_value
=
gru_value
.
output_value
;
}
}
blas
.
GEMM_FREE
(
packed_gate
);
blas
.
GEMM_FREE
(
packed_state
);
math
::
Batch2LoDTensorFunctor
<
DeviceContext
,
T
>
to_seq
;
math
::
Batch2LoDTensorFunctor
<
DeviceContext
,
T
>
to_seq
;
batch_hidden
->
set_lod
(
batch_gate
->
lod
());
batch_hidden
->
set_lod
(
batch_gate
->
lod
());
...
...
paddle/fluid/operators/math/blas.h
浏览文件 @
8c23f7c4
...
@@ -165,7 +165,7 @@ class BlasT : private Blas<DeviceContext> {
...
@@ -165,7 +165,7 @@ class BlasT : private Blas<DeviceContext> {
template
<
typename
...
ARGS
>
template
<
typename
...
ARGS
>
T
*
GEMM_ALLOC
(
ARGS
...
args
)
const
{
T
*
GEMM_ALLOC
(
ARGS
...
args
)
const
{
Base
()
->
template
GEMM_ALLOC
<
T
>(
args
...);
return
Base
()
->
template
GEMM_ALLOC
<
T
>(
args
...);
}
}
template
<
typename
...
ARGS
>
template
<
typename
...
ARGS
>
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录