From c47e258ea489a7773eb6a257b969195dec7642a5 Mon Sep 17 00:00:00 2001 From: baojun <32073718+baojun-nervana@users.noreply.github.com> Date: Wed, 13 Feb 2019 06:06:04 -0800 Subject: [PATCH] Add ngraph sum, sigmoid, relu_grad and tanh_grad op (#15642) * Added ngraph sum op test=develop * Added sigmoid, relu_grad and tanh_grad test=develop * remove duplicates test=develop --- .../fluid/operators/ngraph/ngraph_bridge.cc | 4 ++ paddle/fluid/operators/ngraph/ngraph_ops.h | 2 + .../operators/ngraph/ops/activation_op.h | 52 ++++++++++++++++++ paddle/fluid/operators/ngraph/ops/sum_op.h | 55 +++++++++++++++++++ .../ngraph/test_activation_ngraph_op.py | 12 +--- .../unittests/ngraph/test_sum_ngraph_op.py | 19 +++++++ 6 files changed, 133 insertions(+), 11 deletions(-) create mode 100644 paddle/fluid/operators/ngraph/ops/activation_op.h create mode 100644 paddle/fluid/operators/ngraph/ops/sum_op.h create mode 100644 python/paddle/fluid/tests/unittests/ngraph/test_sum_ngraph_op.py diff --git a/paddle/fluid/operators/ngraph/ngraph_bridge.cc b/paddle/fluid/operators/ngraph/ngraph_bridge.cc index e8b92fc02a5..08d72a5b397 100644 --- a/paddle/fluid/operators/ngraph/ngraph_bridge.cc +++ b/paddle/fluid/operators/ngraph/ngraph_bridge.cc @@ -48,8 +48,12 @@ std::map}, + {"sum", NG_OPS::BuildSumNode}, {"relu", NG_OPS::BuildUnaryNode}, + {"relu_grad", NG_OPS::BuildReluGradNode}, {"tanh", NG_OPS::BuildUnaryNode}, + {"tanh_grad", NG_OPS::BuildTanhGradNode}, {"top_k", NG_OPS::BuildTopKNode}}; void NgraphBridge::BuildNgNode( diff --git a/paddle/fluid/operators/ngraph/ngraph_ops.h b/paddle/fluid/operators/ngraph/ngraph_ops.h index 438a9c1be9e..c7d7392080c 100644 --- a/paddle/fluid/operators/ngraph/ngraph_ops.h +++ b/paddle/fluid/operators/ngraph/ngraph_ops.h @@ -22,6 +22,7 @@ limitations under the License. */ #pragma once #include "ops/accuracy_op.h" +#include "ops/activation_op.h" #include "ops/batch_norm_op.h" #include "ops/binary_unary_op.h" #include "ops/conv2d_op.h" @@ -32,4 +33,5 @@ limitations under the License. */ #include "ops/pool2d_op.h" #include "ops/scale_op.h" #include "ops/softmax_op.h" +#include "ops/sum_op.h" #include "ops/top_k_op.h" diff --git a/paddle/fluid/operators/ngraph/ops/activation_op.h b/paddle/fluid/operators/ngraph/ops/activation_op.h new file mode 100644 index 00000000000..f66080e3aab --- /dev/null +++ b/paddle/fluid/operators/ngraph/ops/activation_op.h @@ -0,0 +1,52 @@ +/*Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. */ + +#pragma once + +#include + +#include "ngraph/ngraph.hpp" +#include "paddle/fluid/platform/ngraph_helper.h" + +namespace paddle { +namespace operators { +namespace ngraphs { + +void BuildReluGradNode( + const std::shared_ptr& op, + std::shared_ptr< + std::unordered_map>> + ngb_node_map) { + auto out = platform::GetInputNode(op, "Out", ngb_node_map); + auto dout = platform::GetInputNode(op, "Out@GRAD", ngb_node_map); + auto relu_grad = std::make_shared(out, dout); + platform::SetOutputNode(op, "X@GRAD", relu_grad, ngb_node_map); +} + +void BuildTanhGradNode( + const std::shared_ptr& op, + std::shared_ptr< + std::unordered_map>> + ngb_node_map) { + auto out = platform::GetInputNode(op, "Out", ngb_node_map); + auto dout = platform::GetInputNode(op, "Out@GRAD", ngb_node_map); + auto shape = out->get_shape(); + auto node_const = + ngraph::op::Constant::create(ngraph::element::f32, shape, {1}); + auto result = dout * (node_const - out * out); + platform::SetOutputNode(op, "X@GRAD", result, ngb_node_map); +} +} // namespace ngraphs +} // namespace operators +} // namespace paddle diff --git a/paddle/fluid/operators/ngraph/ops/sum_op.h b/paddle/fluid/operators/ngraph/ops/sum_op.h new file mode 100644 index 00000000000..97f4ce64aa5 --- /dev/null +++ b/paddle/fluid/operators/ngraph/ops/sum_op.h @@ -0,0 +1,55 @@ +/*Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. + +Licensed under the Apache License, Version 2.0 (the "License"); +you may not use this file except in compliance with the License. +You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + +Unless required by applicable law or agreed to in writing, software +distributed under the License is distributed on an "AS IS" BASIS, +WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +See the License for the specific language governing permissions and +limitations under the License. */ + +#pragma once + +#include +#include + +#include "ngraph/ngraph.hpp" +#include "paddle/fluid/platform/ngraph_helper.h" + +namespace paddle { +namespace operators { +namespace ngraphs { + +void BuildSumNode( + const std::shared_ptr& op, + std::shared_ptr< + std::unordered_map>> + ngb_node_map) { + std::vector op_inputs; + for (auto& var_name_item : op->Inputs()) { + for (auto& var_name : var_name_item.second) { + op_inputs.push_back(var_name); + if (ngb_node_map->find(var_name) == ngb_node_map->end()) { + PADDLE_THROW("op % input varname %s is not found in var_node_map", + op->Type(), var_name); + } + } + } + std::shared_ptr& sum = ngb_node_map->at(op_inputs[0]); + for (size_t k = 1; k < op_inputs.size(); ++k) { + std::shared_ptr& nodek = ngb_node_map->at(op_inputs[k]); + if (nodek->get_element_type() != sum->get_element_type()) { + nodek = + std::make_shared(nodek, sum->get_element_type()); + } + sum = sum + nodek; + } + platform::SetOutputNode(op, "Out", sum, ngb_node_map); +} +} // namespace ngraphs +} // namespace operators +} // namespace paddle diff --git a/python/paddle/fluid/tests/unittests/ngraph/test_activation_ngraph_op.py b/python/paddle/fluid/tests/unittests/ngraph/test_activation_ngraph_op.py index 2bd9bf84303..034d7792c13 100644 --- a/python/paddle/fluid/tests/unittests/ngraph/test_activation_ngraph_op.py +++ b/python/paddle/fluid/tests/unittests/ngraph/test_activation_ngraph_op.py @@ -18,17 +18,7 @@ import unittest import numpy as np import paddle.fluid.core as core from paddle.fluid.tests.unittests.op_test import OpTest -from paddle.fluid.tests.unittests.test_activation_op import TestRelu, TestTanh - - -class TestNGRAPHReluDim2(TestRelu): - def setUp(self): - super(TestNGRAPHReluDim2, self).setUp() - - -class TestNGRAPHTanhDim2(TestTanh): - def setUp(self): - super(TestNGRAPHTanhDim2, self).setUp() +from paddle.fluid.tests.unittests.test_activation_op import TestSigmoid, TestRelu, TestTanh class TestNGRAPHReluDim4(TestRelu): diff --git a/python/paddle/fluid/tests/unittests/ngraph/test_sum_ngraph_op.py b/python/paddle/fluid/tests/unittests/ngraph/test_sum_ngraph_op.py new file mode 100644 index 00000000000..ed9fb618024 --- /dev/null +++ b/python/paddle/fluid/tests/unittests/ngraph/test_sum_ngraph_op.py @@ -0,0 +1,19 @@ +# Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +from __future__ import print_function +import unittest +from paddle.fluid.tests.unittests.test_sum_op import TestSumOp, TestSelectedRowsSumOp, TestLoDTensorAndSelectedRowsOp + +if __name__ == "__main__": + unittest.main() -- GitLab