Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
Paddle
提交
8c0ea4bf
P
Paddle
项目概览
PaddlePaddle
/
Paddle
1 年多 前同步成功
通知
2302
Star
20931
Fork
5422
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1423
列表
看板
标记
里程碑
合并请求
543
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1,423
Issue
1,423
列表
看板
标记
里程碑
合并请求
543
合并请求
543
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
8c0ea4bf
编写于
11月 20, 2020
作者:
J
joanna.wozna.intel
提交者:
GitHub
11月 20, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Add bf16 matmul, fc, elementwise add and mul (#28729)
* Add bf16 matmul, fc, elementwise add and mul * Correct unit test
上级
efc3b182
变更
12
显示空白变更内容
内联
并排
Showing
12 changed file
with
373 addition
and
15 deletion
+373
-15
paddle/fluid/framework/ir/graph_pattern_detector.cc
paddle/fluid/framework/ir/graph_pattern_detector.cc
+8
-3
paddle/fluid/framework/ir/mkldnn/cpu_bfloat16_placement_pass_tester.cc
...framework/ir/mkldnn/cpu_bfloat16_placement_pass_tester.cc
+3
-1
paddle/fluid/operators/elementwise/mkldnn/elementwise_add_mkldnn_op.cc
...operators/elementwise/mkldnn/elementwise_add_mkldnn_op.cc
+2
-0
paddle/fluid/operators/elementwise/mkldnn/elementwise_mul_mkldnn_op.cc
...operators/elementwise/mkldnn/elementwise_mul_mkldnn_op.cc
+2
-0
paddle/fluid/operators/mkldnn/fc_mkldnn_op.cc
paddle/fluid/operators/mkldnn/fc_mkldnn_op.cc
+10
-1
paddle/fluid/operators/mkldnn/matmul_mkldnn_op.cc
paddle/fluid/operators/mkldnn/matmul_mkldnn_op.cc
+16
-3
python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_cpu_bfloat16_pass.py
...s/unittests/ir/inference/test_mkldnn_cpu_bfloat16_pass.py
+2
-7
python/paddle/fluid/tests/unittests/mkldnn/test_elementwise_add_bf16_mkldnn_op.py
...s/unittests/mkldnn/test_elementwise_add_bf16_mkldnn_op.py
+60
-0
python/paddle/fluid/tests/unittests/mkldnn/test_elementwise_mul_bf16_mkldnn_op.py
...s/unittests/mkldnn/test_elementwise_mul_bf16_mkldnn_op.py
+60
-0
python/paddle/fluid/tests/unittests/mkldnn/test_fc_bf16_mkldnn_op.py
...le/fluid/tests/unittests/mkldnn/test_fc_bf16_mkldnn_op.py
+85
-0
python/paddle/fluid/tests/unittests/mkldnn/test_matmul_bf16_mkldnn_op.py
...luid/tests/unittests/mkldnn/test_matmul_bf16_mkldnn_op.py
+121
-0
tools/static_mode_white_list.py
tools/static_mode_white_list.py
+4
-0
未找到文件。
paddle/fluid/framework/ir/graph_pattern_detector.cc
浏览文件 @
8c0ea4bf
...
...
@@ -2101,13 +2101,18 @@ PDNode *patterns::QuantizePlacement::operator()(
PDNode
*
patterns
::
Bfloat16Placement
::
operator
()(
const
std
::
unordered_set
<
std
::
string
>
&
bfloat16_enabled_op_types
)
{
std
::
unordered_set
<
std
::
string
>
supported_op_types
=
std
::
unordered_set
<
std
::
string
>
({
"concat"
,
"conv2d"
,
"fusion_gru"
,
"gelu"
,
"layer_norm"
,
"reshape2"
,
"softmax"
,
std
::
unordered_set
<
std
::
string
>
(
{
"concat"
,
"conv2d"
,
"elementwise_add"
,
"elementwise_mul"
,
"fc"
,
"fusion_gru"
,
"gelu"
,
"layer_norm"
,
"matmul"
,
"reshape2"
,
"softmax"
,
"sum"
,
"transpose2"
});
if
(
!
bfloat16_enabled_op_types
.
empty
())
{
supported_op_types
=
bfloat16_enabled_op_types
;
}
auto
*
op
=
pattern
->
NewNode
(
op_repr
())
->
assert_is_ops
(
supported_op_types
);
op
->
assert_more
([
&
](
Node
*
node
)
{
return
node
->
Op
()
->
GetAttrIfExists
<
bool
>
(
"use_mkldnn"
)
||
node
->
Op
()
->
Type
()
==
"reshape2"
;
});
return
op
;
}
...
...
paddle/fluid/framework/ir/mkldnn/cpu_bfloat16_placement_pass_tester.cc
浏览文件 @
8c0ea4bf
...
...
@@ -24,10 +24,12 @@ namespace ir {
void
SetOp
(
ProgramDesc
*
prog
,
const
std
::
string
&
type
,
const
std
::
string
&
name
,
const
std
::
vector
<
std
::
string
>&
inputs
,
const
std
::
vector
<
std
::
string
>&
outputs
,
const
std
::
string
&
mkldnn_data_type
=
"float32"
)
{
const
std
::
string
&
mkldnn_data_type
=
"float32"
,
const
bool
use_mkldnn
=
true
)
{
auto
*
op
=
prog
->
MutableBlock
(
0
)
->
AppendOp
();
op
->
SetType
(
type
);
if
(
type
!=
"reshape2"
)
op
->
SetAttr
(
"use_mkldnn"
,
use_mkldnn
);
op
->
SetAttr
(
"mkldnn_data_type"
,
mkldnn_data_type
);
if
(
type
==
"conv2d"
)
{
...
...
paddle/fluid/operators/elementwise/mkldnn/elementwise_add_mkldnn_op.cc
浏览文件 @
8c0ea4bf
...
...
@@ -66,6 +66,8 @@ namespace ops = paddle::operators;
REGISTER_OP_KERNEL
(
elementwise_add
,
MKLDNN
,
::
paddle
::
platform
::
CPUPlace
,
ops
::
EltwiseMKLDNNKernel
<
float
,
dnnl
::
algorithm
::
binary_add
>
,
ops
::
EltwiseMKLDNNKernel
<
paddle
::
platform
::
bfloat16
,
dnnl
::
algorithm
::
binary_add
>
,
ops
::
EltwiseMKLDNNKernel
<
int8_t
,
dnnl
::
algorithm
::
binary_add
>
,
ops
::
EltwiseMKLDNNKernel
<
uint8_t
,
dnnl
::
algorithm
::
binary_add
>
)
...
...
paddle/fluid/operators/elementwise/mkldnn/elementwise_mul_mkldnn_op.cc
浏览文件 @
8c0ea4bf
...
...
@@ -19,5 +19,7 @@ namespace ops = paddle::operators;
REGISTER_OP_KERNEL
(
elementwise_mul
,
MKLDNN
,
::
paddle
::
platform
::
CPUPlace
,
ops
::
EltwiseMKLDNNKernel
<
float
,
dnnl
::
algorithm
::
binary_mul
>
,
ops
::
EltwiseMKLDNNKernel
<
paddle
::
platform
::
bfloat16
,
dnnl
::
algorithm
::
binary_mul
>
,
ops
::
EltwiseMKLDNNKernel
<
int8_t
,
dnnl
::
algorithm
::
binary_mul
>
,
ops
::
EltwiseMKLDNNKernel
<
uint8_t
,
dnnl
::
algorithm
::
binary_mul
>
)
paddle/fluid/operators/mkldnn/fc_mkldnn_op.cc
浏览文件 @
8c0ea4bf
...
...
@@ -536,9 +536,13 @@ static void ExecuteFc(const ExecutionContext& ctx, const LoDTensor* input,
framework
::
vectorize
<
int
>
(
w
->
dims
()),
ctx
.
OutputName
(
"Out"
));
constexpr
bool
is_int8
=
std
::
is_same
<
T_in
,
int8_t
>::
value
||
std
::
is_same
<
T_in
,
uint8_t
>::
value
;
if
(
!
is_int8
||
force_fp32_output
)
{
bool
is_bfloat16
=
std
::
is_same
<
T_in
,
paddle
::
platform
::
bfloat16
>::
value
;
if
((
!
is_int8
&&
!
is_bfloat16
)
||
force_fp32_output
)
{
GetPrimitiveFactory
<
T_in
,
T_w
,
float
>
(
dev_ctx
,
prim_key
)
->
ExecuteFcPrimitive
(
input
,
w
,
bias
,
output
,
dev_ctx
,
ctx
);
}
else
if
(
is_bfloat16
)
{
GetPrimitiveFactory
<
T_in
,
T_w
,
platform
::
bfloat16
>
(
dev_ctx
,
prim_key
)
->
ExecuteFcPrimitive
(
input
,
w
,
bias
,
output
,
dev_ctx
,
ctx
);
}
else
if
(
fuse_relu
)
{
GetPrimitiveFactory
<
T_in
,
T_w
,
uint8_t
>
(
dev_ctx
,
prim_key
)
->
ExecuteFcPrimitive
(
input
,
w
,
bias
,
output
,
dev_ctx
,
ctx
);
...
...
@@ -580,6 +584,11 @@ REGISTER_OP_KERNEL_WITH_CUSTOM_TYPE(fc, MKLDNN, ::paddle::platform::CPUPlace,
FP32
,
ops
::
kFCMKLDNNFP32
,
ops
::
FCMKLDNNOpKernel
<
float
,
float
>
);
REGISTER_OP_KERNEL_WITH_CUSTOM_TYPE
(
fc
,
MKLDNN
,
::
paddle
::
platform
::
CPUPlace
,
BF16
,
ops
::
kFCMKLDNNFP32
,
ops
::
FCMKLDNNOpKernel
<
paddle
::
platform
::
bfloat16
,
paddle
::
platform
::
bfloat16
>
);
REGISTER_OP_KERNEL_WITH_CUSTOM_TYPE
(
fc
,
MKLDNN
,
::
paddle
::
platform
::
CPUPlace
,
U8
,
ops
::
kFCMKLDNNINT8
,
ops
::
FCMKLDNNOpKernel
<
uint8_t
,
int8_t
>
);
...
...
paddle/fluid/operators/mkldnn/matmul_mkldnn_op.cc
浏览文件 @
8c0ea4bf
...
...
@@ -42,6 +42,11 @@ constexpr bool IsInt8() {
return
std
::
is_same
<
T
,
int8_t
>::
value
||
std
::
is_same
<
T
,
uint8_t
>::
value
;
}
template
<
typename
T
>
constexpr
bool
IsBfloat16
()
{
return
std
::
is_same
<
T
,
paddle
::
platform
::
bfloat16
>::
value
;
}
// Get row matrix shape from a vector shape. If the rank of x_dim > 1, the
// original x_dim is returned.
static
framework
::
DDim
RowMatrixDimsFromVector
(
const
framework
::
DDim
&
x_dim
)
{
...
...
@@ -170,7 +175,9 @@ class MatMulFactory {
void
CorrectStridesWhenFloatOutputFused
(
const
ExecutionContext
&
ctx
,
const
memory
::
dim
N
,
memory
::
dim
b
,
memory
::
dims
*
out_strides
)
const
{
if
(
!
IsInt8
<
OT
>
()
&&
IsOutputFused
(
ctx
))
*
out_strides
=
{
N
,
b
*
N
,
1
};
if
(
!
IsInt8
<
OT
>
()
&&
!
IsBfloat16
<
OT
>
()
&&
IsOutputFused
(
ctx
))
{
*
out_strides
=
{
N
,
b
*
N
,
1
};
}
}
MatMulDims
GetMatmulDims
(
const
ExecutionContext
&
ctx
)
{
...
...
@@ -348,10 +355,14 @@ static std::shared_ptr<MatMulFactory<XT, YT, OT>> GetPrimitiveFactory(
template
<
typename
XT
,
typename
YT
>
static
void
ExecuteMatMul
(
const
ExecutionContext
&
ctx
)
{
constexpr
bool
is_int8
=
IsInt8
<
XT
>
();
constexpr
bool
is_bfloat16
=
IsBfloat16
<
XT
>
();
const
bool
force_fp32_output
=
ctx
.
Attr
<
bool
>
(
"force_fp32_output"
);
constexpr
bool
fuse_relu
=
false
;
// TODO(intel): Enable eltwise fuses
if
(
!
is_int8
||
force_fp32_output
)
{
if
(
force_fp32_output
||
((
!
is_int8
)
&&
(
!
is_bfloat16
))
)
{
GetPrimitiveFactory
<
XT
,
YT
,
float
>
(
ctx
)
->
CreateAndExecute
(
ctx
);
}
else
if
(
is_bfloat16
)
{
GetPrimitiveFactory
<
XT
,
YT
,
paddle
::
platform
::
bfloat16
>
(
ctx
)
->
CreateAndExecute
(
ctx
);
}
else
if
(
fuse_relu
)
{
GetPrimitiveFactory
<
XT
,
YT
,
uint8_t
>
(
ctx
)
->
CreateAndExecute
(
ctx
);
}
else
{
...
...
@@ -376,5 +387,7 @@ class DNNLMatMulKernel : public framework::OpKernel<T> {
namespace
ops
=
paddle
::
operators
;
REGISTER_OP_KERNEL
(
matmul
,
MKLDNN
,
::
paddle
::
platform
::
CPUPlace
,
ops
::
DNNLMatMulKernel
<
float
>
,
ops
::
DNNLMatMulKernel
<
int8_t
>
,
ops
::
DNNLMatMulKernel
<
float
>
,
ops
::
DNNLMatMulKernel
<
paddle
::
platform
::
bfloat16
>
,
ops
::
DNNLMatMulKernel
<
int8_t
>
,
ops
::
DNNLMatMulKernel
<
uint8_t
>
);
python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_cpu_bfloat16_pass.py
浏览文件 @
8c0ea4bf
...
...
@@ -25,18 +25,13 @@ class TestMKLDNNCpuBfloat16Pass(InferencePassTest):
with
fluid
.
program_guard
(
self
.
main_program
,
self
.
startup_program
):
x
=
fluid
.
data
(
name
=
'x'
,
shape
=
[
-
1
]
+
self
.
shape_x
,
dtype
=
self
.
d_type
)
y
=
fluid
.
data
(
name
=
'y'
,
shape
=
[
-
1
]
+
self
.
shape_y
,
dtype
=
self
.
d_type
)
out
=
fluid
.
layers
.
matmul
(
x
,
y
)
out
=
fluid
.
layers
.
transpose
(
out
,
perm
=
[
0
,
1
,
2
,
3
])
out
=
fluid
.
layers
.
transpose
(
x
,
perm
=
[
0
,
1
,
2
,
3
])
out
=
fluid
.
layers
.
reshape
(
out
,
[
0
,
0
,
0
,
0
])
out
=
fluid
.
layers
.
fc
(
out
,
size
=
1
)
self
.
feeds
=
{
"x"
:
np
.
random
.
random
([
self
.
bs
]
+
self
.
shape_x
).
astype
(
self
.
d_type
),
"y"
:
np
.
random
.
random
([
self
.
bs
]
+
self
.
shape_y
).
astype
(
self
.
d_type
)
np
.
random
.
random
([
self
.
bs
]
+
self
.
shape_x
).
astype
(
self
.
d_type
)
}
self
.
fetch_list
=
[
out
]
...
...
python/paddle/fluid/tests/unittests/mkldnn/test_elementwise_add_bf16_mkldnn_op.py
0 → 100644
浏览文件 @
8c0ea4bf
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
import
unittest
import
numpy
as
np
import
paddle.fluid.core
as
core
from
paddle.fluid.tests.unittests.op_test
import
OpTest
,
convert_float_to_uint16
from
paddle
import
enable_static
@
unittest
.
skipIf
(
not
core
.
supports_bfloat16
(),
"place does not support BF16 evaluation"
)
class
TestElementwiseAddBf16MklDNNOp
(
OpTest
):
def
setUp
(
self
):
self
.
op_type
=
"elementwise_add"
self
.
use_mkldnn
=
True
self
.
mkldnn_data_type
=
"bfloat16"
self
.
axis
=
-
1
self
.
generate_data
()
self
.
inputs
=
{
'X'
:
convert_float_to_uint16
(
self
.
x
),
'Y'
:
convert_float_to_uint16
(
self
.
y
)
}
self
.
attrs
=
{
'axis'
:
self
.
axis
,
'use_mkldnn'
:
self
.
use_mkldnn
}
self
.
outputs
=
{
'Out'
:
convert_float_to_uint16
(
self
.
out
)}
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
(
100
,
).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
(
100
,
).
astype
(
np
.
float32
)
self
.
out
=
np
.
add
(
self
.
x
,
self
.
y
)
def
test_check_output
(
self
):
self
.
check_output_with_place
(
core
.
CPUPlace
())
def
test_check_grad_normal
(
self
):
pass
def
test_check_grad_ingore_x
(
self
):
pass
def
test_check_grad_ingore_y
(
self
):
pass
if
__name__
==
'__main__'
:
enable_static
()
unittest
.
main
()
python/paddle/fluid/tests/unittests/mkldnn/test_elementwise_mul_bf16_mkldnn_op.py
0 → 100644
浏览文件 @
8c0ea4bf
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
import
unittest
import
numpy
as
np
import
paddle.fluid.core
as
core
from
paddle.fluid.tests.unittests.op_test
import
OpTest
,
convert_float_to_uint16
from
paddle
import
enable_static
@
unittest
.
skipIf
(
not
core
.
supports_bfloat16
(),
"place does not support BF16 evaluation"
)
class
TestElementwiseMulBf16MklDNNOp
(
OpTest
):
def
setUp
(
self
):
self
.
op_type
=
"elementwise_mul"
self
.
use_mkldnn
=
True
self
.
mkldnn_data_type
=
"bfloat16"
self
.
axis
=
-
1
self
.
generate_data
()
self
.
inputs
=
{
'X'
:
convert_float_to_uint16
(
self
.
x
),
'Y'
:
convert_float_to_uint16
(
self
.
y
)
}
self
.
attrs
=
{
'axis'
:
self
.
axis
,
'use_mkldnn'
:
self
.
use_mkldnn
}
self
.
outputs
=
{
'Out'
:
convert_float_to_uint16
(
self
.
out
)}
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
(
100
,
).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
(
100
,
).
astype
(
np
.
float32
)
self
.
out
=
np
.
multiply
(
self
.
x
,
self
.
y
)
def
test_check_output
(
self
):
self
.
check_output_with_place
(
core
.
CPUPlace
())
def
test_check_grad_normal
(
self
):
pass
def
test_check_grad_ingore_x
(
self
):
pass
def
test_check_grad_ingore_y
(
self
):
pass
if
__name__
==
'__main__'
:
enable_static
()
unittest
.
main
()
python/paddle/fluid/tests/unittests/mkldnn/test_fc_bf16_mkldnn_op.py
0 → 100644
浏览文件 @
8c0ea4bf
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
import
unittest
import
numpy
as
np
import
paddle.fluid.core
as
core
from
paddle.fluid.tests.unittests.op_test
import
OpTest
,
convert_float_to_uint16
from
paddle
import
enable_static
def
fully_connected_naive
(
input
,
weights
,
bias_data
):
result
=
np
.
dot
(
input
,
weights
)
+
bias_data
return
result
class
MatrixGenerate
:
def
__init__
(
self
,
mb
,
ic
,
oc
,
h
,
w
):
self
.
input
=
np
.
random
.
random
((
mb
,
ic
*
h
*
w
)).
astype
(
np
.
float32
)
self
.
weights
=
np
.
random
.
random
((
ic
*
h
*
w
,
oc
)).
astype
(
np
.
float32
)
@
unittest
.
skipIf
(
not
core
.
supports_bfloat16
(),
"place does not support BF16 evaluation"
)
class
TestFcBf16MklDNNOp
(
OpTest
):
def
generate_data
(
self
):
self
.
matrix
=
MatrixGenerate
(
1
,
10
,
15
,
3
,
3
)
self
.
bias
=
np
.
random
.
random
(
15
).
astype
(
"float32"
)
def
setUp
(
self
):
self
.
op_type
=
"fc"
self
.
use_mkldnn
=
True
self
.
mkldnn_data_type
=
"bfloat16"
self
.
force_fp32_output
=
False
self
.
generate_data
()
self
.
output
=
fully_connected_naive
(
self
.
matrix
.
input
,
self
.
matrix
.
weights
,
self
.
bias
)
if
not
self
.
force_fp32_output
:
self
.
output
=
convert_float_to_uint16
(
self
.
output
)
self
.
inputs
=
{
'Input'
:
convert_float_to_uint16
(
self
.
matrix
.
input
),
'W'
:
self
.
matrix
.
weights
,
'Bias'
:
self
.
bias
}
self
.
attrs
=
{
'use_mkldnn'
:
self
.
use_mkldnn
,
'force_fp32_output'
:
self
.
force_fp32_output
}
self
.
outputs
=
{
'Out'
:
self
.
output
}
def
test_check_output
(
self
):
self
.
check_output_with_place
(
core
.
CPUPlace
())
def
test_check_grad_normal
(
self
):
pass
def
test_check_grad_no_weight
(
self
):
pass
class
TestFCMKLDNNOp1
(
TestFcBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
matrix
=
MatrixGenerate
(
2
,
15
,
48
,
2
,
2
)
self
.
bias
=
np
.
random
.
random
(
48
).
astype
(
np
.
float32
)
if
__name__
==
"__main__"
:
enable_static
()
unittest
.
main
()
python/paddle/fluid/tests/unittests/mkldnn/test_matmul_bf16_mkldnn_op.py
0 → 100644
浏览文件 @
8c0ea4bf
# Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from
__future__
import
print_function
import
unittest
import
os
import
numpy
as
np
import
paddle.fluid.core
as
core
from
paddle.fluid.tests.unittests.op_test
import
OpTest
,
skip_check_grad_ci
,
convert_float_to_uint16
from
paddle
import
enable_static
@
unittest
.
skipIf
(
not
core
.
supports_bfloat16
(),
"place does not support BF16 evaluation"
)
class
TestMatmulBf16MklDNNOp
(
OpTest
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
25
,
2
,
2
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
25
,
2
,
2
)).
astype
(
np
.
float32
)
self
.
alpha
=
1.0
self
.
out
=
self
.
alpha
*
np
.
matmul
(
self
.
x
,
self
.
y
)
def
set_attributes
(
self
):
self
.
alpha
=
self
.
alpha
if
hasattr
(
self
,
'alpha'
)
else
1.0
self
.
attrs
=
{
'alpha'
:
self
.
alpha
,
"use_mkldnn"
:
self
.
use_mkldnn
,
"mkldnn_data_type"
:
self
.
mkldnn_data_type
,
"force_fp32_output"
:
self
.
force_fp32_output
}
def
setUp
(
self
):
self
.
op_type
=
"matmul"
self
.
use_mkldnn
=
True
self
.
dtype
=
np
.
uint16
self
.
mkldnn_data_type
=
"bfloat16"
self
.
force_fp32_output
=
False
self
.
generate_data
()
self
.
set_attributes
()
if
not
self
.
force_fp32_output
:
self
.
out
=
convert_float_to_uint16
(
self
.
out
)
self
.
outputs
=
{
'Out'
:
self
.
out
}
self
.
x
=
convert_float_to_uint16
(
self
.
x
)
self
.
y
=
convert_float_to_uint16
(
self
.
y
)
self
.
inputs
=
{
'X'
:
self
.
x
,
'Y'
:
self
.
y
}
def
test_check_output
(
self
):
self
.
check_output_with_place
(
core
.
CPUPlace
())
def
test_check_grad
(
self
):
pass
class
TestDnnlMatMulOpAlpha
(
TestMatmulBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
17
,
2
,
3
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
17
,
3
,
2
)).
astype
(
np
.
float32
)
self
.
alpha
=
2.0
self
.
out
=
self
.
alpha
*
np
.
matmul
(
self
.
x
,
self
.
y
)
class
TestDnnlMatMulOp2D
(
TestMatmulBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
9
,
12
)).
astype
(
np
.
float32
)
self
.
out
=
np
.
matmul
(
self
.
x
,
self
.
y
)
class
TestDnnlMatMulOpTransposeX
(
TestMatmulBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
out
=
np
.
matmul
(
np
.
transpose
(
self
.
x
),
self
.
y
)
def
set_attributes
(
self
):
self
.
attrs
=
{
"use_mkldnn"
:
self
.
use_mkldnn
,
"mkldnn_data_type"
:
self
.
mkldnn_data_type
,
'transpose_X'
:
True
}
class
TestDnnlMatMulOpTransposeY
(
TestMatmulBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
out
=
np
.
matmul
(
self
.
x
,
np
.
transpose
(
self
.
y
))
def
set_attributes
(
self
):
self
.
attrs
=
{
"use_mkldnn"
:
self
.
use_mkldnn
,
"mkldnn_data_type"
:
self
.
mkldnn_data_type
,
'transpose_Y'
:
True
}
class
TestMatmulBf16MklDNNForceFp32Output
(
TestMatmulBf16MklDNNOp
):
def
generate_data
(
self
):
self
.
x
=
np
.
random
.
random
((
12
,
9
)).
astype
(
np
.
float32
)
self
.
y
=
np
.
random
.
random
((
9
,
12
)).
astype
(
np
.
float32
)
self
.
force_fp32_output
=
True
self
.
alpha
=
0.5
self
.
out
=
self
.
alpha
*
np
.
matmul
(
self
.
x
,
self
.
y
)
if
__name__
==
"__main__"
:
enable_static
()
unittest
.
main
()
tools/static_mode_white_list.py
浏览文件 @
8c0ea4bf
...
...
@@ -590,13 +590,17 @@ STATIC_MODE_TESTING_LIST = [
'test_conv3d_mkldnn_op'
,
'test_dequantize_mkldnn_op'
,
'test_elementwise_add_mkldnn_op'
,
'test_elementwise_add_bf16_mkldnn_op'
,
'test_elementwise_mul_mkldnn_op'
,
'test_elementwise_mul_bf16_mkldnn_op'
,
'test_fc_mkldnn_op'
,
'test_fc_bf16_mkldnn_op'
,
'test_fusion_gru_int8_mkldnn_op'
,
'test_fusion_gru_mkldnn_op'
,
'test_gaussian_random_mkldnn_op'
,
'test_lrn_mkldnn_op'
,
'test_matmul_mkldnn_op'
,
'test_matmul_bf16_mkldnn_op'
,
'test_mul_int8_mkldnn_op'
,
'test_multi_gru_mkldnn_op'
,
'test_pool2d_int8_mkldnn_op'
,
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录