未验证 提交 c9a88013 编写于 作者: Y yaoxuefeng 提交者: GitHub

enhance error messages of lookup_tale, merge_ids, data_norm (#27619)

* enhance error messages of lookup_tale, merge_ids, data_norm

* fix

* fix error msg in .cu
上级 9cc5603d
...@@ -388,7 +388,8 @@ class DataNormKernel<platform::CPUDeviceContext, T> ...@@ -388,7 +388,8 @@ class DataNormKernel<platform::CPUDeviceContext, T>
break; break;
} }
default: default:
PADDLE_THROW("Unknown storage order: %d", data_layout); PADDLE_THROW(platform::errors::InvalidArgument(
"Unknown storage order: %d", data_layout));
} }
} }
}; };
...@@ -464,7 +465,8 @@ class DataNormGradOp : public framework::OperatorWithKernel { ...@@ -464,7 +465,8 @@ class DataNormGradOp : public framework::OperatorWithKernel {
const framework::ExecutionContext &ctx) const override { const framework::ExecutionContext &ctx) const override {
const auto *var = ctx.InputVar(framework::GradVarName("Y")); const auto *var = ctx.InputVar(framework::GradVarName("Y"));
if (var == nullptr) { if (var == nullptr) {
PADDLE_THROW("can't find Y@GRAD"); PADDLE_THROW(platform::errors::InvalidArgument(
"Y@GRAD can not be found for computation"));
} }
const Tensor *t = nullptr; const Tensor *t = nullptr;
if (var->IsType<Tensor>()) { if (var->IsType<Tensor>()) {
...@@ -473,7 +475,8 @@ class DataNormGradOp : public framework::OperatorWithKernel { ...@@ -473,7 +475,8 @@ class DataNormGradOp : public framework::OperatorWithKernel {
t = &var->Get<LoDTensor>(); t = &var->Get<LoDTensor>();
} }
if (t == nullptr) { if (t == nullptr) {
PADDLE_THROW("can't find Y@GRAD"); PADDLE_THROW(platform::errors::InvalidArgument(
"Y@GRAD can not be found for computation"));
} }
// TODO(pzelazko-intel): enable MKLDNN layout when it's ready // TODO(pzelazko-intel): enable MKLDNN layout when it's ready
...@@ -696,7 +699,8 @@ class DataNormGradKernel<platform::CPUDeviceContext, T> ...@@ -696,7 +699,8 @@ class DataNormGradKernel<platform::CPUDeviceContext, T>
break; break;
} }
default: default:
PADDLE_THROW("Unknown storage order: %s", data_layout_str); PADDLE_THROW(platform::errors::InvalidArgument(
"Unknown storage order: %s", data_layout_str));
} }
} }
}; };
......
...@@ -16,6 +16,7 @@ limitations under the License. */ ...@@ -16,6 +16,7 @@ limitations under the License. */
#include <tuple> #include <tuple>
#include <unordered_map> #include <unordered_map>
#include <utility>
#include <vector> #include <vector>
#include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/tensor_util.h" #include "paddle/fluid/framework/tensor_util.h"
...@@ -30,7 +31,8 @@ class MergeIdsOpKernel : public framework::OpKernel<T> { ...@@ -30,7 +31,8 @@ class MergeIdsOpKernel : public framework::OpKernel<T> {
void Compute(const framework::ExecutionContext &ctx) const override { void Compute(const framework::ExecutionContext &ctx) const override {
auto place = ctx.GetPlace(); auto place = ctx.GetPlace();
if (!platform::is_cpu_place(place)) { if (!platform::is_cpu_place(place)) {
PADDLE_THROW("MergeIds do not support GPU kernel"); PADDLE_THROW(platform::errors::InvalidArgument(
"MergeIds do not support GPU kernel"));
} }
const auto ids = ctx.MultiInput<framework::LoDTensor>("Ids"); const auto ids = ctx.MultiInput<framework::LoDTensor>("Ids");
......
...@@ -161,11 +161,12 @@ class LookupTableGradCUDAKernel : public framework::OpKernel<T> { ...@@ -161,11 +161,12 @@ class LookupTableGradCUDAKernel : public framework::OpKernel<T> {
auto d_output_dims_2d = auto d_output_dims_2d =
framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1); framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1);
PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d, PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d,
"ShapeError: The shape of lookup_table@Grad and " platform::errors::InvalidArgument(
"output@Grad should be same. " "ShapeError: The shape of lookup_table@Grad and "
"But received lookup_table@Grad's shape = [%s], " "output@Grad should be same. "
"output@Grad's shape = [%s].", "But received lookup_table@Grad's shape = [%s], "
d_table_value->dims(), d_output_dims_2d); "output@Grad's shape = [%s].",
d_table_value->dims(), d_output_dims_2d));
memory::Copy(gpu_place, d_table_data, gpu_place, d_output_data, memory::Copy(gpu_place, d_table_data, gpu_place, d_output_data,
d_output->numel() * sizeof(T), stream); d_output->numel() * sizeof(T), stream);
......
...@@ -159,9 +159,9 @@ class LookupTableGradKernel : public framework::OpKernel<T> { ...@@ -159,9 +159,9 @@ class LookupTableGradKernel : public framework::OpKernel<T> {
auto *table_t = context.Input<SelectedRows>("W"); auto *table_t = context.Input<SelectedRows>("W");
table_dim = table_t->value().dims(); table_dim = table_t->value().dims();
} else { } else {
PADDLE_THROW( PADDLE_THROW(platform::errors::InvalidArgument(
"The parameter W of a LookupTable " "The parameter W of a LookupTable "
"must be either LoDTensor or SelectedRows"); "must be either LoDTensor or SelectedRows"));
} }
int64_t padding_idx = context.Attr<int64_t>("padding_idx"); int64_t padding_idx = context.Attr<int64_t>("padding_idx");
......
...@@ -28,11 +28,15 @@ class LookupTableV2Op : public framework::OperatorWithKernel { ...@@ -28,11 +28,15 @@ class LookupTableV2Op : public framework::OperatorWithKernel {
void InferShape(framework::InferShapeContext* ctx) const override { void InferShape(framework::InferShapeContext* ctx) const override {
PADDLE_ENFORCE_EQ(ctx->HasInput("W"), true, PADDLE_ENFORCE_EQ(ctx->HasInput("W"), true,
"Input(W) of LookupTableV2Op should not be null."); platform::errors::InvalidArgument(
"Input(W) of LookupTableV2Op should not be null."));
PADDLE_ENFORCE_EQ(ctx->HasInput("Ids"), true, PADDLE_ENFORCE_EQ(ctx->HasInput("Ids"), true,
"Input(Ids) of LookupTableV2Op should not be null."); platform::errors::InvalidArgument(
PADDLE_ENFORCE_EQ(ctx->HasOutput("Out"), true, "Input(Ids) of LookupTableV2Op should not be null."));
"Output(Out) of LookupTableV2Op should not be null."); PADDLE_ENFORCE_EQ(
ctx->HasOutput("Out"), true,
platform::errors::InvalidArgument(
"Output(Out) of LookupTableV2Op should not be null."));
auto table_dims = ctx->GetInputDim("W"); auto table_dims = ctx->GetInputDim("W");
auto ids_dims = ctx->GetInputDim("Ids"); auto ids_dims = ctx->GetInputDim("Ids");
...@@ -40,10 +44,11 @@ class LookupTableV2Op : public framework::OperatorWithKernel { ...@@ -40,10 +44,11 @@ class LookupTableV2Op : public framework::OperatorWithKernel {
VLOG(5) << "ids rank is " << ids_rank << std::endl; VLOG(5) << "ids rank is " << ids_rank << std::endl;
PADDLE_ENFORCE_EQ( PADDLE_ENFORCE_EQ(
table_dims.size(), 2, table_dims.size(), 2,
"ShapeError: The dimensions of the 'lookup table' must be 2. " platform::errors::InvalidArgument(
"But received lookup table's dimensions = %d, " "ShapeError: The dimensions of the 'lookup table' must be 2. "
"lookup table's shape = [%s].", "But received lookup table's dimensions = %d, "
table_dims.size(), table_dims); "lookup table's shape = [%s].",
table_dims.size(), table_dims));
auto output_dims = framework::vectorize(ids_dims); auto output_dims = framework::vectorize(ids_dims);
output_dims.push_back(table_dims[1]); output_dims.push_back(table_dims[1]);
......
...@@ -191,11 +191,12 @@ class LookupTableV2GradCUDAKernel : public framework::OpKernel<T> { ...@@ -191,11 +191,12 @@ class LookupTableV2GradCUDAKernel : public framework::OpKernel<T> {
auto d_output_dims_2d = auto d_output_dims_2d =
framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1); framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1);
PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d, PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d,
"ShapeError: The shape of lookup_table@Grad and " platform::errors::InvalidArgument(
"output@Grad should be same. " "ShapeError: The shape of lookup_table@Grad and "
"But received lookup_table@Grad's shape = [%s], " "output@Grad should be same. "
"output@Grad's shape = [%s].", "But received lookup_table@Grad's shape = [%s], "
d_table_value->dims(), d_output_dims_2d); "output@Grad's shape = [%s].",
d_table_value->dims(), d_output_dims_2d));
memory::Copy(gpu_place, d_table_data, gpu_place, d_output_data, memory::Copy(gpu_place, d_table_data, gpu_place, d_output_data,
d_output->numel() * sizeof(T), stream); d_output->numel() * sizeof(T), stream);
......
...@@ -74,16 +74,18 @@ class LookupTableV2Kernel : public framework::OpKernel<T> { ...@@ -74,16 +74,18 @@ class LookupTableV2Kernel : public framework::OpKernel<T> {
} else { } else {
PADDLE_ENFORCE_LT( PADDLE_ENFORCE_LT(
ids[i], row_number, ids[i], row_number,
"Variable value (input) of OP(fluid.layers.embedding) " platform::errors::InvalidArgument(
"expected >= 0 and < %ld, but got %ld. Please check input " "Variable value (input) of OP(fluid.layers.embedding) "
"value.", "expected >= 0 and < %ld, but got %ld. Please check input "
row_number, ids[i]); "value.",
row_number, ids[i]));
PADDLE_ENFORCE_GE( PADDLE_ENFORCE_GE(
ids[i], 0, ids[i], 0,
"Variable value (input) of OP(fluid.layers.embedding) " platform::errors::InvalidArgument(
"expected >= 0 and < %ld, but got %ld. Please check input " "Variable value (input) of OP(fluid.layers.embedding) "
"value.", "expected >= 0 and < %ld, but got %ld. Please check input "
row_number, ids[i]); "value.",
row_number, ids[i]));
memcpy(output + i * row_width, table + ids[i] * row_width, memcpy(output + i * row_width, table + ids[i] * row_width,
row_width * sizeof(T)); row_width * sizeof(T));
} }
...@@ -101,13 +103,16 @@ class LookupTableV2Kernel : public framework::OpKernel<T> { ...@@ -101,13 +103,16 @@ class LookupTableV2Kernel : public framework::OpKernel<T> {
} else { } else {
PADDLE_ENFORCE_GE( PADDLE_ENFORCE_GE(
ids[i], 0, ids[i], 0,
"Variable value (input) of OP(fluid.layers.embedding) " platform::errors::InvalidArgument(
"expected >= 0. But received %ld", "Variable value (input) of OP(fluid.layers.embedding) "
ids[i]); "expected >= 0. But received %ld",
ids[i]));
auto id_index = table_t.Index(ids[i]); auto id_index = table_t.Index(ids[i]);
PADDLE_ENFORCE_GE(id_index, 0, PADDLE_ENFORCE_GE(
"the input key should be exists. But received %d.", id_index, 0,
id_index); platform::errors::InvalidArgument(
"the input key should be exists. But received %d.",
id_index));
blas.VCOPY(row_width, table + id_index * row_width, blas.VCOPY(row_width, table + id_index * row_width,
output + i * row_width); output + i * row_width);
} }
...@@ -128,9 +133,9 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> { ...@@ -128,9 +133,9 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> {
auto *table_t = context.Input<SelectedRows>("W"); auto *table_t = context.Input<SelectedRows>("W");
table_dim = table_t->value().dims(); table_dim = table_t->value().dims();
} else { } else {
PADDLE_THROW( PADDLE_THROW(platform::errors::InvalidArgument(
"The parameter W of a LookupTableV2 " "The parameter W of a LookupTableV2 "
"must be either LoDTensor or SelectedRows"); "must be either LoDTensor or SelectedRows"));
} }
int64_t padding_idx = context.Attr<int64_t>("padding_idx"); int64_t padding_idx = context.Attr<int64_t>("padding_idx");
...@@ -170,11 +175,12 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> { ...@@ -170,11 +175,12 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> {
auto d_output_dims_2d = auto d_output_dims_2d =
framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1); framework::flatten_to_2d(d_output_dims, d_output_dims.size() - 1);
PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d, PADDLE_ENFORCE_EQ(d_table_value->dims(), d_output_dims_2d,
"ShapeError: The shape of lookup_table@Grad and " platform::errors::InvalidArgument(
"output@Grad should be same. " "ShapeError: The shape of lookup_table@Grad and "
"But received lookup_table@Grad's shape = [%s], " "output@Grad should be same. "
"output@Grad's shape = [%s].", "But received lookup_table@Grad's shape = [%s], "
d_table_value->dims(), d_output_dims_2d); "output@Grad's shape = [%s].",
d_table_value->dims(), d_output_dims_2d));
memcpy(d_table_data, d_output_data, sizeof(T) * d_output->numel()); memcpy(d_table_data, d_output_data, sizeof(T) * d_output->numel());
} else { } else {
...@@ -211,14 +217,18 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> { ...@@ -211,14 +217,18 @@ class LookupTableV2GradKernel : public framework::OpKernel<T> {
} else { } else {
PADDLE_ENFORCE_LT( PADDLE_ENFORCE_LT(
ids_data[i], N, ids_data[i], N,
"Variable value (input) of OP(fluid.layers.embedding) " platform::errors::InvalidArgument(
"expected >= 0 and < %ld, but got %ld. Please check input value.", "Variable value (input) of OP(fluid.layers.embedding) "
N, ids_data[i]); "expected >= 0 and < %ld, but got %ld. Please check input "
"value.",
N, ids_data[i]));
PADDLE_ENFORCE_GE( PADDLE_ENFORCE_GE(
ids_data[i], 0, ids_data[i], 0,
"Variable value (input) of OP(fluid.layers.embedding) " platform::errors::InvalidArgument(
"expected >= 0 and < %ld, but got %ld. Please check input value.", "Variable value (input) of OP(fluid.layers.embedding) "
N, ids_data[i]); "expected >= 0 and < %ld, but got %ld. Please check input "
"value.",
N, ids_data[i]));
for (int j = 0; j < D; ++j) { for (int j = 0; j < D; ++j) {
d_table_data[ids_data[i] * D + j] += d_output_data[i * D + j]; d_table_data[ids_data[i] * D + j] += d_output_data[i * D + j];
} }
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册