未验证 提交 393a0b16 编写于 作者: L Leo Chen 提交者: GitHub

[NPU] refine nan check (#34508)

上级 a6f55e48
...@@ -123,32 +123,30 @@ class CAllReduceOpCPUKernel : public framework::OpKernel<T> { ...@@ -123,32 +123,30 @@ class CAllReduceOpCPUKernel : public framework::OpKernel<T> {
#if defined(PADDLE_WITH_ASCEND_CL) #if defined(PADDLE_WITH_ASCEND_CL)
// return true if found_inf_or_nan or return false; // return true if found_inf_or_nan or return false;
template <typename T> template <typename T>
bool CheckNumerics(const framework::ExecutionContext& exe_ctx, bool ContainsNan(const framework::ExecutionContext& exe_ctx, aclrtStream stream,
aclrtStream stream, const paddle::framework::Tensor* in) { const paddle::framework::Tensor* in) {
auto& dev_ctx = auto& dev_ctx =
exe_ctx.template device_context<paddle::platform::NPUDeviceContext>(); exe_ctx.template device_context<paddle::platform::NPUDeviceContext>();
using Tensor = paddle::framework::Tensor; using Tensor = paddle::framework::Tensor;
Tensor out(in->type()); Tensor out(in->type());
out.Resize(in->dims());
out.mutable_data<T>(dev_ctx.GetPlace()); Tensor mean(in->type());
mean.Resize({1});
bool found_inf_data = false; mean.mutable_data<T>(dev_ctx.GetPlace());
std::vector<int> axes;
try { for (int i = 0; i < in->dims().size(); ++i) {
const auto& runner = axes.push_back(i);
NpuOpRunner("CheckNumerics", {*in}, {out},
{{"message", std::string("check_numberics")}});
runner.Run(stream);
dev_ctx.Wait();
} catch (platform::EnforceNotMet& exception) {
LOG(WARNING) << "[check_nan_and_inf] detected contains NaN or INF!!!";
found_inf_data = true;
} catch (...) {
LOG(WARNING) << "[check_nan_and_inf] detected contains NaN or INF!!!";
found_inf_data = true;
} }
const auto& runner_mean = NpuOpRunner("ReduceMeanD", {*in}, {mean},
{{"axes", axes}, {"keep_dims", false}});
std::vector<T> vec;
TensorToVector(mean, exe_ctx.device_context(), &vec);
return found_inf_data; if (std::isnan(static_cast<float>(vec[0]))) {
return true;
}
return false;
} }
#endif #endif
...@@ -216,22 +214,22 @@ class CAllReduceOpASCENDKernel : public framework::OpKernel<T> { ...@@ -216,22 +214,22 @@ class CAllReduceOpASCENDKernel : public framework::OpKernel<T> {
framework::Tensor tmp; framework::Tensor tmp;
tmp.mutable_data<float>({8}, ctx.GetPlace()); tmp.mutable_data<float>({8}, ctx.GetPlace());
bool check_numerics = false; bool has_nan = false;
auto d_type = in->type(); auto d_type = in->type();
switch (d_type) { switch (d_type) {
case framework::proto::VarType::FP16: case framework::proto::VarType::FP16:
case framework::proto::VarType::FP32: { case framework::proto::VarType::FP32: {
VLOG(4) << "prepare to FoundNanInf"; VLOG(4) << "prepare to check nan";
check_numerics = CheckNumerics<T>(ctx, dev_ctx->stream(), in); has_nan = ContainsNan<T>(ctx, dev_ctx->stream(), in);
VLOG(4) << "check_numerics:" << check_numerics; VLOG(4) << "ContainsNan:" << has_nan;
break; break;
} }
default: default:
break; break;
} }
if (check_numerics) { if (has_nan) {
T inf = static_cast<T>(std::numeric_limits<float>::infinity()); T inf = static_cast<T>(std::numeric_limits<float>::infinity());
VLOG(4) << "fill input data constant inf"; VLOG(4) << "fill input data constant inf";
auto dims = in->dims(); auto dims = in->dims();
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册