/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include #include "../test_helper.h" #include "../test_include.h" #include "operators/math/gemm.h" #include "operators/math/math_function.h" #define a(i, j) a[(i)*lda + (j)] #define b(i, j) b[(i)*ldb + (j)] #define c1(i, j) c1[(i)*ldc + (j)] #define m 1024 #define n 1024 #define k 1024 int main() { paddle_mobile::PaddleMobile paddle_mobile; paddle_mobile.SetThreadNum(4); Tensor aa, bb, cc; auto aaptr = aa.mutable_data({m, k}); auto bbptr = bb.mutable_data({k, n}); auto ccptr = cc.mutable_data({m, n}); for (int i = 0; i < m * k; ++i) { aaptr[i] = 2; } for (int i = 0; i < k * n; ++i) { bbptr[i] = 2; } for (int i = 0; i < m * n; ++i) { ccptr[i] = 2; } Tensor aa_int8, bb_int8, cc_int32, cc_int8; auto aaptr_int8 = aa_int8.mutable_data({m, k}); auto bbptr_int8 = bb_int8.mutable_data({k, n}); auto ccptr_int32 = cc_int32.mutable_data({m, n}); auto ccptr_int8 = cc_int8.mutable_data({m, n}); int32_t* bias_data = new int32_t[m]; for (int i = 0; i < m * k; ++i) { aaptr_int8[i] = static_cast(2); } for (int i = 0; i < k * n; ++i) { bbptr_int8[i] = static_cast(2); } for (int i = 0; i < m * n; ++i) { ccptr_int32[i] = static_cast(2); } for (int i = 0; i < m; ++i) { bias_data[i] = 2; } // float // warm-up 10 times for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul( aa, false, bb, false, static_cast(1), &cc, static_cast(0), false, nullptr); } auto time1 = time(); for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul( aa, false, bb, false, static_cast(1), &cc, static_cast(0), false, nullptr); } auto time2 = time(); std::cout << "float gemm cost :" << time_diff(time1, time2) / 10 << "ms\n"; // int8_t without bias // warm-up 10 times for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul_int8( aa_int8, false, bb_int8, false, static_cast(1), &cc_int32, static_cast(0), false, nullptr); } auto time3 = time(); for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul_int8( aa_int8, false, bb_int8, false, static_cast(1), &cc_int32, static_cast(0), false, nullptr); } auto time4 = time(); std::cout << "int8_t gemm cost :" << time_diff(time3, time4) / 10 << "ms\n"; // int8_t with bias&relu // warm-up 10 times for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul_int8( aa_int8, false, bb_int8, false, static_cast(0.618), &cc_int8, static_cast(0), true, &bias_data[0]); } auto time5 = time(); for (int j = 0; j < 10; ++j) { paddle_mobile::operators::math::matmul_int8( aa_int8, false, bb_int8, false, static_cast(0.618), &cc_int8, static_cast(0), true, &bias_data[0]); } auto time6 = time(); std::cout << "int8_t gemm_with_bias_relu cost :" << time_diff(time5, time6) / 10 << "ms\n"; delete[] bias_data; return 0; }