// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include "lite/core/kernel.h" #include "lite/core/op_lite.h" #include "lite/core/op_registry.h" #include "lite/fluid/eigen.h" #include "lite/operators/activation_ops.h" namespace paddle { namespace lite { namespace kernels { namespace x86 { enum ActBwdOpFwdDeps { kNoDeps = 0x00, // Do not need any forward input/output kDepX = 0x01, // Only need forward input X kDepOut = 0x02, // Only need forward output Out // Never add kDepXOut, because Out can be always calculated // by forward input X in backward part. // FIXME(zjl): but in MKLDNN abs, X and Out are all needed... // Developers should not rely on this enum value! kDepXOut = 0x03 }; template struct BaseActivationFunctor { using ELEMENT_TYPE = T; using AttrPair = std::vector>; AttrPair GetAttrs() { return AttrPair(); } /* NOTE(*): Output reuse X memory if X is not dependented by its Gradient. For example, sigmoid op's gradient didn't involve x, so its output can reuse input memory. But abs op's gradient use x, it can not be inplaced. gradient did use x. */ bool Inplace() const { return false; } }; template bool Activate(const lite::Tensor* X, lite::Tensor* Out) { using T = typename Functor::ELEMENT_TYPE; auto place = lite::fluid::EigenDeviceType(); CHECK_OR_FALSE(X) CHECK_OR_FALSE(Out) auto x = lite::fluid::EigenVector::Flatten(*X); auto out = lite::fluid::EigenVector::Flatten(*Out); Functor()(place, x, out); return true; } // square(x) = x^2 template struct SquareFunctor : public BaseActivationFunctor { template void operator()(Device d, X x, Out out) const { out.device(d) = x.square(); } }; template class SquareCompute : public KernelLite { public: using param_t = operators::ActivationParam; void Run() override { auto& param = *param_.get_mutable(); param.Out->template mutable_data(); Activate>(param.X, param.Out); } virtual ~SquareCompute() = default; }; // relu(x) = max(x, 0) template struct ReluFunctor : public BaseActivationFunctor { template void operator()(Device d, X x, Out out) const { out.device(d) = x.cwiseMax(static_cast(0)); } }; template class ReluCompute : public KernelLite { public: using param_t = operators::ActivationParam; void Run() override { auto& param = *param_.get_mutable(); param.Out->template mutable_data(); Activate>(param.X, param.Out); } virtual ~ReluCompute() = default; }; // softsign(x) = x / (1 + |x|) template struct SoftsignFunctor : public BaseActivationFunctor { template void operator()(Device d, X x, Out out) { out.device(d) = x / (static_cast(1) + x.abs()); } }; template class SoftsignCompute : public KernelLite { public: using param_t = operators::ActivationParam; void Run() override { // auto& context = ctx_->As(); auto& param = *param_.get_mutable(); param.Out->template mutable_data(); Activate>(param.X, param.Out); } virtual ~SoftsignCompute() = default; }; } // namespace x86 } // namespace kernels } // namespace lite } // namespace paddle