/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include "paddle/fluid/operators/math/jit_kernel.h" #include #include // for memcpy #include #include #include "gflags/gflags.h" #include "glog/logging.h" #include "gtest/gtest.h" #ifdef PADDLE_WITH_MKLML #include "paddle/fluid/platform/dynload/mklml.h" #endif #ifdef __AVX__ #include #endif constexpr int repeat = 20000; inline double GetCurrentUS() { struct timeval time; gettimeofday(&time, NULL); return 1e+6 * time.tv_sec + time.tv_usec; } template void RandomVec(const int n, T* a, const T lower = static_cast(-20.f), const T upper = static_cast(20.f)) { static unsigned int seed = 100; std::mt19937 rng(seed++); std::uniform_real_distribution uniform_dist(0, 1); for (int i = 0; i < n; ++i) { a[i] = static_cast(uniform_dist(rng) * (upper - lower) + lower); } } void vaddbias_ref(const int n, const float a, const float* x, float* y) { for (int i = 0; i < n; ++i) { y[i] = x[i] + a; } } TEST(JitKernel, vaddbias) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 64, 100, 128, 256}) { std::vector x(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data(), -2.f, 2.f); const auto& ker = jit::KernelPool::Instance().template Get>(d); const float a = 2.f; const float* x_data = x.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vaddbias_ref(d, a, x_data, zref_data); } auto trefe = GetCurrentUS(); auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, a, x_data, ztgt_data); } auto ttgte = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat << " us, tgt takes: " << (ttgte - ttgts) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } void vexp_ref(const int n, const float* x, float* y) { for (int i = 0; i < n; ++i) { y[i] = std::exp(x[i]); } } #ifdef PADDLE_WITH_MKLML void vexp_mkl(const int n, const float* x, float* y) { paddle::platform::dynload::vsExp(n, x, y); } #endif TEST(JitKernel, vexp) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 128}) { std::vector x(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data(), -2.f, 2.f); const auto& ker = jit::KernelPool::Instance().template Get>(d); const float* x_data = x.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vexp_ref(d, x_data, zref_data); } auto trefe = GetCurrentUS(); #ifdef PADDLE_WITH_MKLML auto tmkls = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vexp_mkl(d, x_data, zref_data); } auto tmkle = GetCurrentUS(); #endif auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, x_data, ztgt_data); } auto ttgte = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat #ifdef PADDLE_WITH_MKLML << " us, mkl takes: " << (tmkle - tmkls) / repeat << " us, " #else << " us, " #endif << "tgt takes: " << (ttgte - ttgts) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } inline float _sigmoid(float x) { const float min = SIGMOID_THRESHOLD_MIN; const float max = SIGMOID_THRESHOLD_MAX; float tmp = (x < min) ? min : ((x > max) ? max : x); return 1.f / (1.f + std::exp(-tmp)); } void vsigmoid_ref(const int n, const float* x, float* y) { for (int i = 0; i < n; ++i) { y[i] = _sigmoid(x[i]); } } void vsigmoid_better( const std::shared_ptr< const paddle::operators::math::jitkernel::VExpKernel>& vexp, const int n, const float* x, float* y) { const float min = SIGMOID_THRESHOLD_MIN; const float max = SIGMOID_THRESHOLD_MAX; for (int i = 0; i < n; ++i) { y[i] = (x[i] < min) ? min : ((x[i] > max) ? max : x[i]); y[i] = 0.f - y[i]; } vexp->Compute(n, y, y); for (int i = 0; i < n; ++i) { y[i] = 1.f / (1.f + y[i]); } } TEST(JitKernel, vsigmoid) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 32, 64, 100, 128, 256}) { std::vector x(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data(), -2.f, 2.f); const auto& ker = jit::KernelPool::Instance().template Get>(d); const auto& vexp = jit::KernelPool::Instance().template Get>(d); const float* x_data = x.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto tmkls = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vsigmoid_better(vexp, d, x_data, zref_data); } auto tmkle = GetCurrentUS(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vsigmoid_ref(d, x_data, zref_data); } auto trefe = GetCurrentUS(); auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, x_data, ztgt_data); } auto ttgte = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat << " us, better(jit exp) takes: " << (tmkle - tmkls) / repeat << " us, tgt takes: " << (ttgte - ttgts) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } void vscal_ref(const int n, const float a, const float* x, float* y) { for (int i = 0; i < n; ++i) { y[i] = a * x[i]; } } void vscal_inp_ref(const int n, const float a, float* x) { for (int i = 0; i < n; ++i) { x[i] = a * x[i]; } } #if defined __AVX__ || defined __AVX2__ void vscal_intri8(const int n, const float a, const float* x, float* y) { __m256 tmp; __m256 scalar = _mm256_set1_ps(a); tmp = _mm256_loadu_ps(x); tmp = _mm256_mul_ps(tmp, scalar); _mm256_storeu_ps(y, tmp); } void vscal_inp_intri8(const int n, const float a, float* x) { __m256 tmp; __m256 scalar = _mm256_set1_ps(a); tmp = _mm256_loadu_ps(x); tmp = _mm256_mul_ps(tmp, scalar); _mm256_storeu_ps(x, tmp); } #endif #ifdef PADDLE_WITH_MKLML void vscal_inp_mkl(const int n, const float a, float* x) { paddle::platform::dynload::cblas_sscal(n, a, x, 1); } #endif TEST(JitKernel, vscal) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 256, 512}) { std::vector x(d), y(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data()); std::memcpy(y.data(), x.data(), sizeof(float) * d); float a = 2.f; const auto& ker = jit::KernelPool::Instance().template Get>(d); const float* x_data = x.data(); float* y_data = y.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vscal_ref(d, a, x_data, zref_data); } auto trefe = GetCurrentUS(); auto trefs1 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vscal_inp_ref(d, a, y_data); } auto trefe1 = GetCurrentUS(); #ifdef PADDLE_WITH_MKLML auto tmkls = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vscal_inp_mkl(d, a, y_data); } auto tmkle = GetCurrentUS(); #endif #if defined __AVX__ || defined __AVX2__ if (d == 8) { auto si0 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vscal_intri8(d, a, x_data, zref_data); } auto si1 = GetCurrentUS(); auto si2 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vscal_inp_intri8(d, a, y_data); } auto si3 = GetCurrentUS(); VLOG(3) << "Vec size 8 intr takes: " << (si1 - si0) / repeat << " us, inplace: " << (si3 - si2) / repeat; } #endif auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, a, x_data, ztgt_data); } auto ttgte = GetCurrentUS(); auto ttgts1 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, a, y_data); } auto ttgte1 = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat << " us, inplace takes: " << (trefe1 - trefs1) / repeat #ifdef PADDLE_WITH_MKLML << " us, mkl inplace takes: " << (tmkle - tmkls) / repeat << " us, " #else << " us, " #endif << "tgt takes: " << (ttgte - ttgts) / repeat << "us, tgt inplace takes: " << (ttgte1 - ttgts1) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } void vmul_ref(const int n, const float* x, const float* y, float* z) { for (int i = 0; i < n; ++i) { z[i] = x[i] * y[i]; } } #if defined __AVX__ || defined __AVX2__ void vmul_intri8(const int n, const float* x, const float* y, float* z) { __m256 tmpx, tmpy; tmpx = _mm256_loadu_ps(x); tmpy = _mm256_loadu_ps(y); tmpx = _mm256_mul_ps(tmpx, tmpy); _mm256_storeu_ps(z, tmpx); } #endif #ifdef PADDLE_WITH_MKLML void vmul_mkl(const int n, const float* x, const float* y, float* z) { paddle::platform::dynload::vsMul(n, x, y, z); } #endif TEST(JitKernel, vmul) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 256, 512}) { std::vector x(d), y(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data()); RandomVec(d, y.data()); const auto& ker = jit::KernelPool::Instance().template Get>(d); const float* x_data = x.data(); const float* y_data = y.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vmul_ref(d, x_data, y_data, zref_data); } auto trefe = GetCurrentUS(); #ifdef PADDLE_WITH_MKLML auto tmkls = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vmul_mkl(d, x_data, y_data, zref_data); } auto tmkle = GetCurrentUS(); #endif #if defined __AVX__ || defined __AVX2__ if (d == 8) { auto si0 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vmul_intri8(d, x_data, y_data, zref_data); } auto si1 = GetCurrentUS(); VLOG(3) << "Vec size 8 intr takes: " << (si1 - si0) / repeat; } #endif auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, x_data, y_data, ztgt_data); } auto ttgte = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat #ifdef PADDLE_WITH_MKLML << " us, mkl takes: " << (tmkle - tmkls) / repeat << " us, " #else << " us, " #endif << "tgt takes: " << (ttgte - ttgts) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } void vadd_ref(const int n, const float* x, const float* y, float* z) { for (int i = 0; i < n; ++i) { z[i] = x[i] + y[i]; } } #if defined __AVX__ || defined __AVX2__ void vadd_intri8(const int n, const float* x, const float* y, float* z) { __m256 tmpx, tmpy; tmpx = _mm256_loadu_ps(x); tmpy = _mm256_loadu_ps(y); tmpx = _mm256_add_ps(tmpx, tmpy); _mm256_storeu_ps(z, tmpx); } #endif #ifdef PADDLE_WITH_MKLML void vadd_mkl(const int n, const float* x, const float* y, float* z) { paddle::platform::dynload::vsAdd(n, x, y, z); } #endif TEST(JitKernel, vadd) { namespace jit = paddle::operators::math::jitkernel; for (int d : {7, 8, 15, 16, 30, 256, 512}) { std::vector x(d), y(d); std::vector zref(d), ztgt(d); RandomVec(d, x.data()); RandomVec(d, y.data()); const auto& ker = jit::KernelPool::Instance().template Get>(d); const float* x_data = x.data(); const float* y_data = y.data(); float* ztgt_data = ztgt.data(); float* zref_data = zref.data(); auto trefs = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vadd_ref(d, x_data, y_data, zref_data); } auto trefe = GetCurrentUS(); #ifdef PADDLE_WITH_MKLML auto tmkls = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vadd_mkl(d, x_data, y_data, zref_data); } auto tmkle = GetCurrentUS(); #endif #if defined __AVX__ || defined __AVX2__ if (d == 8) { auto si0 = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { vadd_intri8(d, x_data, y_data, zref_data); } auto si1 = GetCurrentUS(); VLOG(3) << "Vec size 8 intr takes: " << (si1 - si0) / repeat; } #endif auto ttgts = GetCurrentUS(); for (int i = 0; i < repeat; ++i) { ker->Compute(d, x_data, y_data, ztgt_data); } auto ttgte = GetCurrentUS(); VLOG(3) << "Vec size " << d << ": refer takes: " << (trefe - trefs) / repeat #ifdef PADDLE_WITH_MKLML << " us, mkl takes: " << (tmkle - tmkls) / repeat << " us, " #else << " us, " #endif << "tgt takes: " << (ttgte - ttgts) / repeat; for (int i = 0; i < d; ++i) { EXPECT_NEAR(ztgt_data[i], zref_data[i], 1e-3); } } } TEST(JitKernel, pool) { namespace jit = paddle::operators::math::jitkernel; const int frame_size = 4; std::string act_gate = "sigmoid", act_cand = "tanh", act_cell = "tanh"; const auto& plstm1 = jit::KernelPool::Instance() .template Get, int, const std::string&, const std::string&, const std::string&>( frame_size, act_gate, act_cand, act_cell); const auto& plstm2 = jit::KernelPool::Instance() .template Get, int, const std::string&, const std::string&, const std::string&>( frame_size, act_gate, act_cand, act_cell); EXPECT_EQ(plstm1, plstm2); const auto& pvmul_f = jit::KernelPool::Instance().template Get>(4); EXPECT_TRUE(std::dynamic_pointer_cast(plstm2) != std::dynamic_pointer_cast(pvmul_f)); const auto& pvmul_d = jit::KernelPool::Instance().template Get>(4); EXPECT_TRUE(std::dynamic_pointer_cast(pvmul_f) != std::dynamic_pointer_cast(pvmul_d)); const auto& pvmul_from_key = jit::KernelPool::Instance().Get("vmulf4"); EXPECT_EQ(pvmul_f, pvmul_from_key); const auto& pvmul_from_key2 = jit::KernelPool::Instance().Get("vmulf5"); EXPECT_TRUE(pvmul_from_key2 == nullptr); }