Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
ad81f22c
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
ad81f22c
编写于
3月 16, 2022
作者:
Q
qipengh
提交者:
GitHub
3月 16, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[MLU] support amp O1 of mlu (#40461)
上级
f748b433
变更
6
显示空白变更内容
内联
并排
Showing
6 changed file
with
45 addition
and
13 deletion
+45
-13
paddle/fluid/framework/data_device_transform.cc
paddle/fluid/framework/data_device_transform.cc
+8
-0
paddle/fluid/imperative/amp_auto_cast.cc
paddle/fluid/imperative/amp_auto_cast.cc
+11
-1
paddle/fluid/operators/batch_norm_op_mlu.cc
paddle/fluid/operators/batch_norm_op_mlu.cc
+14
-9
python/paddle/fluid/contrib/mixed_precision/fp16_lists.py
python/paddle/fluid/contrib/mixed_precision/fp16_lists.py
+3
-0
python/paddle/fluid/dygraph/amp/auto_cast.py
python/paddle/fluid/dygraph/amp/auto_cast.py
+7
-2
python/paddle/fluid/dygraph/amp/loss_scaler.py
python/paddle/fluid/dygraph/amp/loss_scaler.py
+2
-1
未找到文件。
paddle/fluid/framework/data_device_transform.cc
浏览文件 @
ad81f22c
...
@@ -34,6 +34,14 @@ void TransDataDevice(const Tensor &in, const platform::Place &dst_place,
...
@@ -34,6 +34,14 @@ void TransDataDevice(const Tensor &in, const platform::Place &dst_place,
return
;
return
;
}
}
// NOTE(hqp): Special case for CPU->MLU, avoid stream sync.
if
(
platform
::
is_cpu_place
(
in
.
place
())
&&
platform
::
is_mlu_place
(
dst_place
))
{
paddle
::
framework
::
TensorCopy
(
in
,
dst_place
,
*
platform
::
DeviceContextPool
::
Instance
().
Get
(
dst_place
),
out
);
return
;
}
// NOTE(yy): TransDataDevice should wait for computation of input.
// NOTE(yy): TransDataDevice should wait for computation of input.
if
(
!
platform
::
is_cuda_pinned_place
(
in
.
place
()))
{
if
(
!
platform
::
is_cuda_pinned_place
(
in
.
place
()))
{
platform
::
DeviceContextPool
::
Instance
().
Get
(
in
.
place
())
->
Wait
();
platform
::
DeviceContextPool
::
Instance
().
Get
(
in
.
place
())
->
Wait
();
...
...
paddle/fluid/imperative/amp_auto_cast.cc
浏览文件 @
ad81f22c
...
@@ -124,7 +124,7 @@ AmpOperators::AmpOperators()
...
@@ -124,7 +124,7 @@ AmpOperators::AmpOperators()
OpSupportedInfos
(
"GPU"
,
paddle
::
framework
::
proto
::
VarType
::
BF16
));
OpSupportedInfos
(
"GPU"
,
paddle
::
framework
::
proto
::
VarType
::
BF16
));
unsupported_bf16_ops_
->
insert
(
unsupported_ops_gpu_bf16
.
begin
(),
unsupported_bf16_ops_
->
insert
(
unsupported_ops_gpu_bf16
.
begin
(),
unsupported_ops_gpu_bf16
.
end
());
unsupported_ops_gpu_bf16
.
end
());
// NOTE: GPU/NPU/XPU is compiled seperatly.
// NOTE: GPU/NPU/XPU
/MLU
is compiled seperatly.
#elif defined(PADDLE_WITH_ASCEND_CL)
#elif defined(PADDLE_WITH_ASCEND_CL)
auto
unsupported_ops_npu_fp16
=
std
::
get
<
2
>
(
auto
unsupported_ops_npu_fp16
=
std
::
get
<
2
>
(
OpSupportedInfos
(
"NPU"
,
paddle
::
framework
::
proto
::
VarType
::
FP16
));
OpSupportedInfos
(
"NPU"
,
paddle
::
framework
::
proto
::
VarType
::
FP16
));
...
@@ -143,6 +143,15 @@ AmpOperators::AmpOperators()
...
@@ -143,6 +143,15 @@ AmpOperators::AmpOperators()
OpSupportedInfos
(
"XPU"
,
paddle
::
framework
::
proto
::
VarType
::
BF16
));
OpSupportedInfos
(
"XPU"
,
paddle
::
framework
::
proto
::
VarType
::
BF16
));
unsupported_bf16_ops_
->
insert
(
unsupported_ops_xpu_bf16
.
begin
(),
unsupported_bf16_ops_
->
insert
(
unsupported_ops_xpu_bf16
.
begin
(),
unsupported_ops_xpu_bf16
.
end
());
unsupported_ops_xpu_bf16
.
end
());
#elif defined(PADDLE_WITH_MLU)
auto
unsupported_ops_mlu_fp16
=
std
::
get
<
2
>
(
OpSupportedInfos
(
"MLU"
,
paddle
::
framework
::
proto
::
VarType
::
FP16
));
unsupported_fp16_ops_
->
insert
(
unsupported_ops_mlu_fp16
.
begin
(),
unsupported_ops_mlu_fp16
.
end
());
auto
unsupported_ops_mlu_bf16
=
std
::
get
<
2
>
(
OpSupportedInfos
(
"MLU"
,
paddle
::
framework
::
proto
::
VarType
::
BF16
));
unsupported_bf16_ops_
->
insert
(
unsupported_ops_mlu_bf16
.
begin
(),
unsupported_ops_mlu_bf16
.
end
());
#endif
#endif
VLOG
(
4
)
<<
allow_ops_
->
size
()
<<
" "
<<
block_ops_
->
size
()
<<
" "
VLOG
(
4
)
<<
allow_ops_
->
size
()
<<
" "
<<
block_ops_
->
size
()
<<
" "
<<
unsupported_fp16_ops_
->
size
()
<<
" "
<<
unsupported_fp16_ops_
->
size
()
<<
" "
...
@@ -210,6 +219,7 @@ inline bool NeedCast(const std::shared_ptr<VarType>& var) {
...
@@ -210,6 +219,7 @@ inline bool NeedCast(const std::shared_ptr<VarType>& var) {
if
(
paddle
::
platform
::
is_gpu_place
(
place
)
||
if
(
paddle
::
platform
::
is_gpu_place
(
place
)
||
paddle
::
platform
::
is_cuda_pinned_place
(
place
)
||
paddle
::
platform
::
is_cuda_pinned_place
(
place
)
||
paddle
::
platform
::
is_xpu_place
(
place
)
||
paddle
::
platform
::
is_xpu_place
(
place
)
||
paddle
::
platform
::
is_mlu_place
(
place
)
||
paddle
::
platform
::
is_npu_place
(
place
)
||
paddle
::
platform
::
is_npu_place
(
place
)
||
paddle
::
platform
::
is_npu_pinned_place
(
place
))
{
paddle
::
platform
::
is_npu_pinned_place
(
place
))
{
// CudaPinndePlace is added for varbase created by dataloader
// CudaPinndePlace is added for varbase created by dataloader
...
...
paddle/fluid/operators/batch_norm_op_mlu.cc
浏览文件 @
ad81f22c
...
@@ -13,6 +13,7 @@ See the License for the specific language governing permissions and
...
@@ -13,6 +13,7 @@ See the License for the specific language governing permissions and
limitations under the License. */
limitations under the License. */
#include "paddle/fluid/operators/batch_norm_op.h"
#include "paddle/fluid/operators/batch_norm_op.h"
#include "paddle/fluid/operators/amp/fp16_type_traits.h"
#include "paddle/fluid/operators/mlu/mlu_baseop.h"
#include "paddle/fluid/operators/mlu/mlu_baseop.h"
namespace
paddle
{
namespace
paddle
{
...
@@ -20,6 +21,8 @@ namespace operators {
...
@@ -20,6 +21,8 @@ namespace operators {
template
<
typename
T
>
template
<
typename
T
>
class
MLUBatchNormOpKernel
:
public
framework
::
OpKernel
<
T
>
{
class
MLUBatchNormOpKernel
:
public
framework
::
OpKernel
<
T
>
{
using
MPDType
=
typename
details
::
MPTypeTrait
<
T
>::
Type
;
public:
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
const
auto
&
place
=
ctx
.
GetPlace
();
const
auto
&
place
=
ctx
.
GetPlace
();
...
@@ -68,10 +71,10 @@ class MLUBatchNormOpKernel : public framework::OpKernel<T> {
...
@@ -68,10 +71,10 @@ class MLUBatchNormOpKernel : public framework::OpKernel<T> {
// alloc memory
// alloc memory
y
->
mutable_data
<
T
>
(
place
);
y
->
mutable_data
<
T
>
(
place
);
mean_out
->
mutable_data
<
T
>
(
place
);
mean_out
->
mutable_data
<
MPDType
>
(
place
);
variance_out
->
mutable_data
<
T
>
(
place
);
variance_out
->
mutable_data
<
MPDType
>
(
place
);
saved_mean
->
mutable_data
<
T
>
(
place
);
saved_mean
->
mutable_data
<
MPDType
>
(
place
);
saved_variance
->
mutable_data
<
T
>
(
place
);
saved_variance
->
mutable_data
<
MPDType
>
(
place
);
Tensor
transformed_x
;
Tensor
transformed_x
;
Tensor
transformed_y
;
Tensor
transformed_y
;
...
@@ -132,6 +135,8 @@ class MLUBatchNormOpKernel : public framework::OpKernel<T> {
...
@@ -132,6 +135,8 @@ class MLUBatchNormOpKernel : public framework::OpKernel<T> {
template
<
typename
T
>
template
<
typename
T
>
class
MLUBatchNormGradOpKernel
:
public
framework
::
OpKernel
<
T
>
{
class
MLUBatchNormGradOpKernel
:
public
framework
::
OpKernel
<
T
>
{
using
MPDType
=
typename
details
::
MPTypeTrait
<
T
>::
Type
;
public:
public:
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
void
Compute
(
const
framework
::
ExecutionContext
&
ctx
)
const
override
{
const
auto
*
x
=
ctx
.
Input
<
Tensor
>
(
"X"
);
const
auto
*
x
=
ctx
.
Input
<
Tensor
>
(
"X"
);
...
@@ -154,10 +159,10 @@ class MLUBatchNormGradOpKernel : public framework::OpKernel<T> {
...
@@ -154,10 +159,10 @@ class MLUBatchNormGradOpKernel : public framework::OpKernel<T> {
auto
&
dev_ctx
=
ctx
.
template
device_context
<
MLUDeviceContext
>();
auto
&
dev_ctx
=
ctx
.
template
device_context
<
MLUDeviceContext
>();
auto
d_x_tmp
=
auto
d_x_tmp
=
ctx
.
AllocateTmpTensor
<
T
,
MLUDeviceContext
>
(
x
->
dims
(),
dev_ctx
);
ctx
.
AllocateTmpTensor
<
T
,
MLUDeviceContext
>
(
x
->
dims
(),
dev_ctx
);
auto
scale_grad_tmp
=
auto
scale_grad_tmp
=
ctx
.
AllocateTmpTensor
<
MPDType
,
MLUDeviceContext
>
(
ctx
.
AllocateTmpTensor
<
T
,
MLUDeviceContext
>
(
scale
->
dims
(),
dev_ctx
);
scale
->
dims
(),
dev_ctx
);
auto
bias_grad_tmp
=
auto
bias_grad_tmp
=
ctx
.
AllocateTmpTensor
<
T
,
MLUDeviceContext
>
(
bias
->
dims
(),
dev_ctx
);
ctx
.
AllocateTmpTensor
<
MPDType
,
MLUDeviceContext
>
(
bias
->
dims
(),
dev_ctx
);
if
(
d_x
==
nullptr
)
{
if
(
d_x
==
nullptr
)
{
d_x
=
&
d_x_tmp
;
d_x
=
&
d_x_tmp
;
...
@@ -171,8 +176,8 @@ class MLUBatchNormGradOpKernel : public framework::OpKernel<T> {
...
@@ -171,8 +176,8 @@ class MLUBatchNormGradOpKernel : public framework::OpKernel<T> {
const
auto
&
place
=
ctx
.
GetPlace
();
const
auto
&
place
=
ctx
.
GetPlace
();
d_x
->
mutable_data
<
T
>
(
place
);
d_x
->
mutable_data
<
T
>
(
place
);
d_scale
->
mutable_data
<
T
>
(
place
);
d_scale
->
mutable_data
<
MPDType
>
(
place
);
d_bias
->
mutable_data
<
T
>
(
place
);
d_bias
->
mutable_data
<
MPDType
>
(
place
);
use_global_stats
=
is_test
||
use_global_stats
;
use_global_stats
=
is_test
||
use_global_stats
;
...
...
python/paddle/fluid/contrib/mixed_precision/fp16_lists.py
浏览文件 @
ad81f22c
...
@@ -173,6 +173,9 @@ if core.is_compiled_with_xpu():
...
@@ -173,6 +173,9 @@ if core.is_compiled_with_xpu():
elif
core
.
is_compiled_with_npu
():
elif
core
.
is_compiled_with_npu
():
_
,
_
,
_sys_unsupported_fp16_list
=
core
.
op_supported_infos
(
_
,
_
,
_sys_unsupported_fp16_list
=
core
.
op_supported_infos
(
'NPU'
,
core
.
VarDesc
.
VarType
.
FP16
)
'NPU'
,
core
.
VarDesc
.
VarType
.
FP16
)
elif
core
.
is_compiled_with_mlu
():
_
,
_
,
_sys_unsupported_fp16_list
=
core
.
op_supported_infos
(
'MLU'
,
core
.
VarDesc
.
VarType
.
FP16
)
else
:
else
:
_
,
_
,
_sys_unsupported_fp16_list
=
core
.
op_supported_infos
(
_
,
_
,
_sys_unsupported_fp16_list
=
core
.
op_supported_infos
(
'GPU'
,
core
.
VarDesc
.
VarType
.
FP16
)
'GPU'
,
core
.
VarDesc
.
VarType
.
FP16
)
...
...
python/paddle/fluid/dygraph/amp/auto_cast.py
浏览文件 @
ad81f22c
...
@@ -271,13 +271,14 @@ def amp_guard(enable=True,
...
@@ -271,13 +271,14 @@ def amp_guard(enable=True,
"current_tracer is None, maybe it is not in imperative mode."
)
"current_tracer is None, maybe it is not in imperative mode."
)
# check device_type:
# check device_type:
# NOTE: Now, amp only support gpu for float16 and bfloat16, xpu for float16, npu for float16.
# NOTE: Now, amp only support gpu for float16 and bfloat16, xpu for float16,
mlu for float16,
npu for float16.
# Maybe we will support cpu for bfloat16.
# Maybe we will support cpu for bfloat16.
if
enable
and
not
(
tracer
.
_expected_place
.
is_gpu_place
()
or
if
enable
and
not
(
tracer
.
_expected_place
.
is_gpu_place
()
or
tracer
.
_expected_place
.
is_xpu_place
()
or
tracer
.
_expected_place
.
is_xpu_place
()
or
tracer
.
_expected_place
.
is_mlu_place
()
or
tracer
.
_expected_place
.
is_npu_place
()):
tracer
.
_expected_place
.
is_npu_place
()):
warnings
.
warn
(
warnings
.
warn
(
'amp_guard can only be enabled on CUDAPlace, XPUPlace, and NPUPlace, current place is %s, so it makes no effect.'
'amp_guard can only be enabled on CUDAPlace, XPUPlace,
MLUPlace,
and NPUPlace, current place is %s, so it makes no effect.'
%
tracer
.
_expected_place
)
%
tracer
.
_expected_place
)
enable
=
False
enable
=
False
# For npu:
# For npu:
...
@@ -288,6 +289,10 @@ def amp_guard(enable=True,
...
@@ -288,6 +289,10 @@ def amp_guard(enable=True,
if
tracer
.
_expected_place
.
is_xpu_place
()
and
(
dtype
==
'bfloat16'
):
if
tracer
.
_expected_place
.
is_xpu_place
()
and
(
dtype
==
'bfloat16'
):
warnings
.
warn
(
'XPUPlace only support float16 amp.'
)
warnings
.
warn
(
'XPUPlace only support float16 amp.'
)
enable
=
False
enable
=
False
# For mlu:
if
tracer
.
_expected_place
.
is_mlu_place
()
and
(
dtype
==
'bfloat16'
):
warnings
.
warn
(
'MLUPlace only support float16 amp.'
)
enable
=
False
# For gpu float16: Compute Capability should >= 7.
# For gpu float16: Compute Capability should >= 7.
# For gpu bfloat16: Compute Capability should >= 8 & CUDA Version should >= 11.
# For gpu bfloat16: Compute Capability should >= 8 & CUDA Version should >= 11.
if
tracer
.
_expected_place
.
is_gpu_place
():
if
tracer
.
_expected_place
.
is_gpu_place
():
...
...
python/paddle/fluid/dygraph/amp/loss_scaler.py
浏览文件 @
ad81f22c
...
@@ -106,9 +106,10 @@ class AmpScaler(object):
...
@@ -106,9 +106,10 @@ class AmpScaler(object):
if
enable
and
not
(
tracer
.
_expected_place
.
is_gpu_place
()
or
if
enable
and
not
(
tracer
.
_expected_place
.
is_gpu_place
()
or
tracer
.
_expected_place
.
is_xpu_place
()
or
tracer
.
_expected_place
.
is_xpu_place
()
or
tracer
.
_expected_place
.
is_mlu_place
()
or
tracer
.
_expected_place
.
is_npu_place
()):
tracer
.
_expected_place
.
is_npu_place
()):
warnings
.
warn
(
warnings
.
warn
(
'AmpScaler can only be enabled on CUDAPlace, XPUPlace and NPUPlace, current place is %s, so it makes no effect.'
'AmpScaler can only be enabled on CUDAPlace, XPUPlace
, MLUPlace
and NPUPlace, current place is %s, so it makes no effect.'
%
tracer
.
_expected_place
)
%
tracer
.
_expected_place
)
enable
=
False
enable
=
False
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录