/** * \file dnn/src/aarch64/matrix_mul/int8_dot/strategy.cpp * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") * * Copyright (c) 2014-2020 Megvii Inc. All rights reserved. * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. */ #include "src/aarch64/matrix_mul/int8_dot/strategy.h" #include "src/aarch64/matrix_mul/asm/common.h" #include "src/arm_common/simd_macro/marm_neon.h" #include "src/common/utils.h" #include "src/aarch64/matrix_mul/int8_dot/kernel_8x12x4.h" #include "src/aarch64/matrix_mul/int8_dot/kernel_mk4_8x12x4.h" #if __ARM_FEATURE_DOTPROD using namespace megdnn; using namespace aarch64; using namespace aarch64::matmul; /* ====================== gemm_s8_8x12 ===========================*/ MEGDNN_REG_GEMM_STRATEGY_IMPL(gemm_s8_8x12); void gemm_s8_8x12::pack_A(dt_int8* outptr, const dt_int8* inptr, int ldin, int y0, int ymax, int k0, int kmax, bool transpose) const { if (transpose) { matmul_8x12x4::gemm_s8_8x12_pack_A_t(outptr, inptr, ldin, y0, ymax, k0, kmax); } else { matmul_8x12x4::gemm_s8_8x12_pack_A_n(outptr, inptr, ldin, y0, ymax, k0, kmax); } } void gemm_s8_8x12::pack_B(dt_int8* out, const dt_int8* in, int ldin, int x0, int xmax, int k0, int kmax, bool transpose) const { if (transpose) { matmul_8x12x4::gemm_s8_8x12_pack_B_t(out, in, ldin, x0, xmax, k0, kmax); } else { matmul_8x12x4::gemm_s8_8x12_pack_B_n(out, in, ldin, x0, xmax, k0, kmax); } } void gemm_s8_8x12::kern(const dt_int8* packA, const dt_int8* packB, size_t M, size_t N, size_t K, dt_int32* C, size_t LDC, bool is_first_k, const dt_int32*, dt_int32*) const { megdnn_assert(A_dtype.enumv() == B_dtype.enumv() && ((A_dtype.enumv() == DTypeEnum::Int8 && C_dtype.enumv() == DTypeEnum::Int32) || (A_dtype.enumv() == DTypeEnum::QuantizedS8 && C_dtype.enumv() == DTypeEnum::QuantizedS32)), "A: %s B: %s C: %s", A_dtype.name(), B_dtype.name(), C_dtype.name()); MEGDNN_MARK_USED_VAR(A_dtype); MEGDNN_MARK_USED_VAR(B_dtype); MEGDNN_MARK_USED_VAR(C_dtype); constexpr size_t A_INTERLEAVE = 8; constexpr size_t B_INTERLEAVE = 12; //! K is packed to times of 4 K = round_up(K, 4); const int K8 = (K << 3); const int K12 = K * 12; const int K4 = K * 4; size_t m = 0; for (; m + A_INTERLEAVE - 1 < M; m += A_INTERLEAVE) { int32_t* output = C + (m * LDC); size_t n = 0; const dt_int8* cur_packB = packB; for (; n + B_INTERLEAVE - 1 < N; n += B_INTERLEAVE) { matmul_8x12x4::kern_8x12(packA, cur_packB, K, output, LDC, is_first_k); output += B_INTERLEAVE; cur_packB += K12; } for (; n < N; n += 4) { matmul_8x12x4::kern_8x4(packA, cur_packB, K, output, LDC, is_first_k, std::min(N - n, 4)); output += 4; cur_packB += K4; } packA += K8; } for (; m < M; m += 4) { int32_t* output = C + (m * LDC); const dt_int8* cur_packB = packB; size_t n = 0; for (; n + B_INTERLEAVE - 1 < N; n += B_INTERLEAVE) { matmul_8x12x4::kern_4x12(packA, cur_packB, K, output, LDC, is_first_k, std::min(M - m, 4)); output += B_INTERLEAVE; cur_packB += K12; } for (; n < N; n += 4) { matmul_8x12x4::kern_4x4(packA, cur_packB, K, output, LDC, is_first_k, std::min(M - m, 4), std::min(N - n, 4)); output += 4; cur_packB += K4; } packA += K4; } } /* ====================== gemm_mk4_s8_8x12 ===========================*/ MEGDNN_REG_GEMM_STRATEGY_IMPL(gemm_mk4_s8_8x12); void gemm_mk4_s8_8x12::pack_A(dt_int8* outptr, const dt_int8* inptr, int ldin, int y0, int ymax, int k0, int kmax, bool transpose) const { megdnn_assert(!transpose, "matrix mul mk4 with transposed matrix A is not supported"); matmul_mk4_8x12x4::gemm_mk4_s8_8x12_pack_A(outptr, inptr, ldin, y0, ymax, k0, kmax); } void gemm_mk4_s8_8x12::pack_B(dt_int8* out, const dt_int8* in, int ldin, int x0, int xmax, int k0, int kmax, bool transpose) const { megdnn_assert(!transpose, "matrix mul mk4 with transposed matrix B is not supported"); matmul_mk4_8x12x4::gemm_mk4_s8_8x12_pack_B(out, in, ldin, x0, xmax, k0, kmax); } void gemm_mk4_s8_8x12::kern(const dt_int8* packA, const dt_int8* packB, size_t M, size_t N, size_t K, dt_int32* C, size_t LDC, bool is_first_k, const dt_int32*, dt_int32*) const { megdnn_assert(A_dtype.enumv() == B_dtype.enumv() && ((A_dtype.enumv() == DTypeEnum::Int8 && C_dtype.enumv() == DTypeEnum::Int32) || (A_dtype.enumv() == DTypeEnum::QuantizedS8 && C_dtype.enumv() == DTypeEnum::QuantizedS32)), "A: %s B: %s C: %s", A_dtype.name(), B_dtype.name(), C_dtype.name()); MEGDNN_MARK_USED_VAR(A_dtype); MEGDNN_MARK_USED_VAR(B_dtype); MEGDNN_MARK_USED_VAR(C_dtype); constexpr size_t A_INTERLEAVE = 8; constexpr size_t B_INTERLEAVE = 12; //! K is packed to times of 4 K = round_up(K, 4); const int K8 = (K << 3); const int K12 = K * 12; const int K4 = K * 4; size_t m = 0; for (; m + A_INTERLEAVE - 1 < M; m += A_INTERLEAVE) { int32_t* output = C + ((m >> 2) * LDC); size_t n = 0; const dt_int8* cur_packB = packB; for (; n + B_INTERLEAVE - 1 < N; n += B_INTERLEAVE) { matmul_mk4_8x12x4::kern_8x12(packA, cur_packB, K, output, LDC, is_first_k); output += (B_INTERLEAVE << 2); cur_packB += K12; } for (; n < N; n += 4) { matmul_mk4_8x12x4::kern_8x4(packA, cur_packB, K, output, LDC, is_first_k, std::min(N - n, 4)); output += 16; cur_packB += K4; } packA += K8; } for (; m < M; m += 4) { int32_t* output = C + ((m >> 2) * LDC); const dt_int8* cur_packB = packB; size_t n = 0; for (; n + B_INTERLEAVE - 1 < N; n += B_INTERLEAVE) { matmul_mk4_8x12x4::kern_4x12(packA, cur_packB, K, output, LDC, is_first_k); output += (B_INTERLEAVE << 2); cur_packB += K12; } for (; n < N; n += 4) { matmul_mk4_8x12x4::kern_4x4(packA, cur_packB, K, output, LDC, is_first_k, std::min(N - n, 4)); output += 16; cur_packB += K4; } packA += K4; } } #endif // vim: syntax=cpp.doxygen