/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #pragma once #include #include "paddle/fluid/operators/jit/kernel_base.h" namespace paddle { namespace operators { namespace jit { namespace more { namespace mix { using T = float; void VSigmoid(const T* x, T* y, int n); void VTanh(const T* x, T* y, int n); void Softmax(const T* x, T* y, int n, int bs); void LSTMCtHt(lstm_t* step, const lstm_attr_t* attr); void LSTMC1H1(lstm_t* step, const lstm_attr_t* attr); void GRUH1(gru_t* step, const gru_attr_t* attr); void GRUHtPart1(gru_t* step, const gru_attr_t* attr); void GRUHtPart2(gru_t* step, const gru_attr_t* attr); #define DECLARE_MORE_KERNEL(name) \ class name##Kernel : public KernelMore> { \ public: \ name##Kernel() { this->func = name; } \ bool UseMe(const typename name##Tuple::attr_type&) const override; \ const char* ImplType() const override { return "Mixed"; } \ } // XYN DECLARE_MORE_KERNEL(VSigmoid); DECLARE_MORE_KERNEL(VTanh); // XRN DECLARE_MORE_KERNEL(Softmax); DECLARE_MORE_KERNEL(LSTMCtHt); DECLARE_MORE_KERNEL(LSTMC1H1); DECLARE_MORE_KERNEL(GRUH1); DECLARE_MORE_KERNEL(GRUHtPart1); DECLARE_MORE_KERNEL(GRUHtPart2); #undef DECLARE_MORE_KERNEL } // namespace mix } // namespace more } // namespace jit } // namespace operators } // namespace paddle