/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "paddle/fluid/operators/jit/more/intrinsic/crf_decoding.h" #include "paddle/fluid/operators/jit/refer/refer.h" #include "paddle/fluid/operators/jit/registry.h" #include "paddle/fluid/platform/cpu_info.h" namespace paddle { namespace operators { namespace jit { namespace more { namespace mkl { template <> void VMul(const float* x, const float* y, float* z, int n) { platform::dynload::vsMul(n, x, y, z); } template <> void VMul(const double* x, const double* y, double* z, int n) { platform::dynload::vdMul(n, x, y, z); } template <> void VAdd(const float* x, const float* y, float* z, int n) { platform::dynload::vsAdd(n, x, y, z); } template <> void VAdd(const double* x, const double* y, double* z, int n) { platform::dynload::vdAdd(n, x, y, z); } template <> void VScal(const float* a, const float* x, float* y, int n) { if (x == y) { platform::dynload::cblas_sscal(n, *a, y, 1); } else { refer::VScal(a, x, y, n); } } template <> void VScal(const double* a, const double* x, double* y, int n) { if (x == y) { platform::dynload::cblas_dscal(n, *a, y, 1); } else { refer::VScal(a, x, y, n); } } template <> void VExp(const float* x, float* y, int n) { platform::dynload::vsExp(n, x, y); } template <> void VExp(const double* x, double* y, int n) { platform::dynload::vdExp(n, x, y); } // TODO(TJ): tuning me carefully on AVX, AVX2 and AVX512 template <> bool VMulKernel::UseMe(int d) const { return platform::MayIUse(platform::avx512f) && d > 512; } template <> bool VAddKernel::UseMe(int d) const { return platform::MayIUse(platform::avx512f) && d > 512; } template <> bool VScalKernel::UseMe(int d) const { return platform::MayIUse(platform::avx512f) && d > 512; } template <> bool VExpKernel::UseMe(int d) const { return d > 7; } template <> bool VSigmoidKernel::UseMe(int d) const { return d > 7; } template <> bool VTanhKernel::UseMe(int d) const { return d > 7; } #define AWALYS_USE_ME_WITH_DOUBLE(func) \ template <> \ bool func##Kernel::UseMe(int d) const { \ return true; \ } AWALYS_USE_ME_WITH_DOUBLE(VMul); AWALYS_USE_ME_WITH_DOUBLE(VAdd); AWALYS_USE_ME_WITH_DOUBLE(VScal); AWALYS_USE_ME_WITH_DOUBLE(VExp); AWALYS_USE_ME_WITH_DOUBLE(VSigmoid); AWALYS_USE_ME_WITH_DOUBLE(VTanh); #undef AWALYS_USE_ME_WITH_DOUBLE } // namespace mkl } // namespace more } // namespace jit } // namespace operators } // namespace paddle namespace mkl = paddle::operators::jit::more::mkl; #define REGISTER_MKL_KERNEL(key, func) \ REGISTER_JITKERNEL_MORE(key, mkl, mkl::func##Kernel, \ mkl::func##Kernel) REGISTER_MKL_KERNEL(vmul, VMul); REGISTER_MKL_KERNEL(vadd, VAdd); REGISTER_MKL_KERNEL(vscal, VScal); REGISTER_MKL_KERNEL(vexp, VExp); REGISTER_MKL_KERNEL(vsigmoid, VSigmoid); REGISTER_MKL_KERNEL(vtanh, VTanh); #undef REGISTER_MKL_KERNEL