“bfc046a0d7b63ca5af2eb44e0dd3f5b11cb3d769”上不存在“git@gitcode.net:paddlepaddle/Serving.git”
mkl.h 3.5 KB
Newer Older
T
tensor-tang 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16
/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License. */

#pragma once

17
#include <cmath>
T
tensor-tang 已提交
18
#include <type_traits>
T
tensor-tang 已提交
19
#include "paddle/fluid/operators/jit/kernel_base.h"
T
tensor-tang 已提交
20 21 22

namespace paddle {
namespace operators {
T
tensor-tang 已提交
23
namespace jit {
T
tensor-tang 已提交
24 25 26
namespace more {
namespace mkl {

27 28 29
template <typename T>
void MatMul(const T* a, const T* b, T* c, int m, int n, int k);

T
tensor-tang 已提交
30 31 32 33
template <typename T>
void VMul(const T* x, const T* y, T* z, int n);

template <typename T>
34 35 36 37 38
void VAdd(const T* x, const T* y, T* z, int n);

template <typename T>
void VScal(const T* a, const T* x, T* y, int n);

39 40 41
template <typename T>
void VExp(const T* x, T* y, int n);

T
tensor-tang 已提交
42 43 44
template <typename T>
void VSquare(const T* x, T* y, int n);

45 46 47 48 49 50
template <typename T>
void VCopy(const T* x, T* y, int n);

template <typename T>
void VAXPY(T a, const T* x, T* y, int n);

51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75
template <typename T>
void VSigmoid(const T* x, T* y, int n) {
  const T min = SIGMOID_THRESHOLD_MIN;
  const T max = SIGMOID_THRESHOLD_MAX;
  for (int i = 0; i < n; ++i) {
    y[i] = (x[i] < min) ? min : ((x[i] > max) ? max : x[i]);
    y[i] = static_cast<T>(0) - y[i];
  }
  VExp(y, y, n);
  for (int i = 0; i < n; ++i) {
    y[i] = static_cast<T>(1) / (static_cast<T>(1) + y[i]);
  }
}

template <typename T>
void VTanh(const T* x, T* y, int n) {
  for (int i = 0; i < n; ++i) {
    y[i] = static_cast<T>(2) * x[i];
  }
  VSigmoid(y, y, n);
  for (int i = 0; i < n; ++i) {
    y[i] = static_cast<T>(2) * y[i] - static_cast<T>(1);
  }
}

76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92
template <typename T>
void SeqPool(const T* x, T* y, const seq_pool_attr_t* attr) {
  VCopy<T>(x, y, attr->w);
  for (int h = 1; h != attr->h; ++h) {
    VAXPY<T>(static_cast<T>(1), x + h * attr->w, y, attr->w);
  }
  if (attr->type == SeqPoolType::kAvg || attr->type == SeqPoolType::kSqrt) {
    T scalar = static_cast<T>(1);
    if (attr->type == SeqPoolType::kAvg) {
      scalar = scalar / static_cast<T>(attr->h);
    } else {
      scalar = scalar / std::sqrt(static_cast<T>(attr->h));
    }
    VScal<T>(&scalar, y, y, attr->w);
  }
}

T
tensor-tang 已提交
93 94 95 96 97 98 99
#define DECLARE_MKL_KERNEL(name, tuples)                             \
  template <typename T>                                              \
  class name##Kernel : public KernelMore<tuples<T>> {                \
   public:                                                           \
    name##Kernel() { this->func = name<T>; }                         \
    bool UseMe(const typename tuples<T>::attr_type&) const override; \
    const char* ImplType() const override { return "MKL"; }          \
T
tensor-tang 已提交
100
  }
101

102 103 104
// ABCMNK
DECLARE_MKL_KERNEL(MatMul, MatMulTuples);

105 106 107 108 109 110 111
// XYZN
DECLARE_MKL_KERNEL(VMul, XYZNTuples);
DECLARE_MKL_KERNEL(VAdd, XYZNTuples);

// AXYN
DECLARE_MKL_KERNEL(VScal, AXYNTuples);

112 113 114 115
// XYN
DECLARE_MKL_KERNEL(VExp, XYNTuples);
DECLARE_MKL_KERNEL(VSigmoid, XYNTuples);
DECLARE_MKL_KERNEL(VTanh, XYNTuples);
T
tensor-tang 已提交
116
DECLARE_MKL_KERNEL(VSquare, XYNTuples);
117

118 119
DECLARE_MKL_KERNEL(SeqPool, SeqPoolTuples);

120
#undef DECLARE_MKL_KERNEL
T
tensor-tang 已提交
121 122 123

}  // namespace mkl
}  // namespace more
T
tensor-tang 已提交
124
}  // namespace jit
T
tensor-tang 已提交
125 126
}  // namespace operators
}  // namespace paddle