From 22f03c3981ae930e608b3e53dcaf32c85408be55 Mon Sep 17 00:00:00 2001 From: dongzhihong Date: Tue, 8 Aug 2017 23:35:48 +0800 Subject: [PATCH] "fix clang format" --- paddle/operators/add_op.cc | 1 + paddle/operators/add_op_test.cc | 1 - paddle/operators/cross_entropy_op.cc | 5 +++-- paddle/operators/mul_op.cc | 1 + paddle/operators/mul_op.cu | 1 + paddle/operators/mul_op.h | 1 + paddle/operators/sigmoid_op.cu | 1 + paddle/operators/softmax_op.cc | 1 + paddle/operators/softmax_op.cu | 1 + 9 files changed, 10 insertions(+), 3 deletions(-) diff --git a/paddle/operators/add_op.cc b/paddle/operators/add_op.cc index 64f856ac6f0..086245ef62d 100644 --- a/paddle/operators/add_op.cc +++ b/paddle/operators/add_op.cc @@ -57,5 +57,6 @@ class AddOpGrad : public framework::OperatorWithKernel { namespace ops = paddle::operators; REGISTER_OP(add_two, ops::AddOp, ops::AddOpMaker); REGISTER_GRADIENT_OP(add_two, add_two_grad, ops::AddOpGrad); + REGISTER_OP_CPU_KERNEL(add_two, ops::AddKernel); diff --git a/paddle/operators/add_op_test.cc b/paddle/operators/add_op_test.cc index 4f33f46bb21..bf529defb20 100644 --- a/paddle/operators/add_op_test.cc +++ b/paddle/operators/add_op_test.cc @@ -17,7 +17,6 @@ limitations under the License. */ #include "paddle/framework/op_registry.h" USE_OP(add_two); -// USE_OP(add_two_grad); TEST(AddOp, GetOpProto) { auto& protos = paddle::framework::OpRegistry::protos(); diff --git a/paddle/operators/cross_entropy_op.cc b/paddle/operators/cross_entropy_op.cc index 7c03fbd9e9e..c813d54e17f 100644 --- a/paddle/operators/cross_entropy_op.cc +++ b/paddle/operators/cross_entropy_op.cc @@ -70,8 +70,9 @@ OnehotCrossEntropy Operator. namespace ops = paddle::operators; REGISTER_OP(onehot_cross_entropy, ops::OnehotCrossEntropyOp, ops::OnehotCrossEntropyOpMaker); -REGISTER_OP_CPU_KERNEL(onehot_cross_entropy, - ops::OnehotCrossEntropyOpKernel); +REGISTER_OP_CPU_KERNEL( + onehot_cross_entropy, + ops::OnehotCrossEntropyOpKernel); REGISTER_GRADIENT_OP(onehot_cross_entropy, onehot_cross_entropy_grad, ops::OnehotCrossEntropyGradientOp); REGISTER_OP_CPU_KERNEL( diff --git a/paddle/operators/mul_op.cc b/paddle/operators/mul_op.cc index 06bc6172db0..db81fd555d1 100644 --- a/paddle/operators/mul_op.cc +++ b/paddle/operators/mul_op.cc @@ -67,4 +67,5 @@ class MulOpGrad : public framework::OperatorWithKernel { namespace ops = paddle::operators; REGISTER_OP(mul, ops::MulOp, ops::MulOpMaker); REGISTER_GRADIENT_OP(mul, mul_grad, ops::MulOpGrad); + REGISTER_OP_CPU_KERNEL(mul, ops::MulKernel); diff --git a/paddle/operators/mul_op.cu b/paddle/operators/mul_op.cu index 346a7e505d1..43debbc21a3 100644 --- a/paddle/operators/mul_op.cu +++ b/paddle/operators/mul_op.cu @@ -16,4 +16,5 @@ #include "paddle/operators/mul_op.h" namespace ops = paddle::operators; + REGISTER_OP_GPU_KERNEL(mul, ops::MulKernel); diff --git a/paddle/operators/mul_op.h b/paddle/operators/mul_op.h index 67c3958149d..ab12631c034 100644 --- a/paddle/operators/mul_op.h +++ b/paddle/operators/mul_op.h @@ -45,5 +45,6 @@ class MulKernel : public framework::OpKernel { Z.device(place) = X.contract(Y, dim_pair); } }; + } // namespace operators } // namespace paddle diff --git a/paddle/operators/sigmoid_op.cu b/paddle/operators/sigmoid_op.cu index 9518c3091ae..1a50dfe14a7 100644 --- a/paddle/operators/sigmoid_op.cu +++ b/paddle/operators/sigmoid_op.cu @@ -16,6 +16,7 @@ #include "paddle/operators/sigmoid_op.h" namespace ops = paddle::operators; + REGISTER_OP_GPU_KERNEL(sigmoid, ops::SigmoidKernel); REGISTER_OP_GPU_KERNEL( diff --git a/paddle/operators/softmax_op.cc b/paddle/operators/softmax_op.cc index 1143fb66c67..3dd4e86918a 100644 --- a/paddle/operators/softmax_op.cc +++ b/paddle/operators/softmax_op.cc @@ -63,6 +63,7 @@ class SoftmaxOpGrad : public framework::OperatorWithKernel { } // namespace paddle namespace ops = paddle::operators; + REGISTER_OP(softmax, ops::SoftmaxOp, ops::SoftmaxOpMaker); REGISTER_OP_CPU_KERNEL(softmax, ops::SoftmaxKernel); diff --git a/paddle/operators/softmax_op.cu b/paddle/operators/softmax_op.cu index 92d22142734..2e99a89699d 100644 --- a/paddle/operators/softmax_op.cu +++ b/paddle/operators/softmax_op.cu @@ -16,6 +16,7 @@ #include "paddle/operators/softmax_op.h" namespace ops = paddle::operators; + REGISTER_OP_GPU_KERNEL(softmax, ops::SoftmaxKernel); REGISTER_OP_GPU_KERNEL( -- GitLab