diff --git a/paddle/operators/norm_op.cc b/paddle/operators/norm_op.cc index 1d9b55d887ed1bd980dc208dc0154c370925890f..b198b76cd49ca7c05b047d42df149d2b1e461b8e 100644 --- a/paddle/operators/norm_op.cc +++ b/paddle/operators/norm_op.cc @@ -55,14 +55,6 @@ class NormOpMaker : public framework::OpProtoAndCheckerMaker { }; class NormOp : public framework::OperatorWithKernel { - protected: - framework::OpKernelType GetKernelType( - const framework::ExecutionContext& ctx) const override { - return framework::OpKernelType( - framework::ToDataType(ctx.Input("X")->type()), - ctx.device_context()); - } - public: using framework::OperatorWithKernel::OperatorWithKernel; void InferShape(framework::InferShapeContext* ctx) const override { @@ -80,14 +72,6 @@ class NormOp : public framework::OperatorWithKernel { }; class NormOpGrad : public framework::OperatorWithKernel { - protected: - framework::OpKernelType GetKernelType( - const framework::ExecutionContext& ctx) const override { - return framework::OpKernelType( - framework::ToDataType(ctx.Input("X")->type()), - ctx.device_context()); - } - public: using framework::OperatorWithKernel::OperatorWithKernel; void InferShape(framework::InferShapeContext* ctx) const override { @@ -105,7 +89,7 @@ REGISTER_OP(norm, ops::NormOp, ops::NormOpMaker, norm_grad, ops::NormOpGrad); REGISTER_OP_CPU_KERNEL( norm, ops::NormKernel, - ops::NormKernel); + ops::NormKernel); REGISTER_OP_CPU_KERNEL( norm_grad, ops::NormGradKernel, - ops::NormGradKernel); + ops::NormGradKernel); diff --git a/paddle/operators/norm_op.cu b/paddle/operators/norm_op.cu index 7d84aaa73248d6bbf1cf104105ce2a2847000eea..2941c89b93141388d3746cf128848960ed6f3625 100644 --- a/paddle/operators/norm_op.cu +++ b/paddle/operators/norm_op.cu @@ -18,7 +18,7 @@ limitations under the License. */ namespace ops = paddle::operators; REGISTER_OP_CUDA_KERNEL( norm, ops::NormKernel, - ops::NormKernel); + ops::NormKernel); REGISTER_OP_CUDA_KERNEL( norm_grad, ops::NormGradKernel, - ops::NormGradKernel); + ops::NormGradKernel);