/** * \file dnn/src/cuda/matrix_mul/algos.h * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") * * Copyright (c) 2014-2020 Megvii Inc. All rights reserved. * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. */ #pragma once #include "megdnn/oprs.h" #include "src/common/utils.h" #include "src/cuda/matrix_mul/opr_impl.h" #include #if CUDA_VERSION >= 10010 #include #endif namespace megdnn { namespace cuda { /*! * \brief base class for matrix mul algos * */ class MatrixMulForwardImpl::AlgoBase : public Algorithm { protected: ~AlgoBase() = default; public: struct SizeArgs { MatrixMulForwardImpl* opr; TensorLayout layout_a, layout_b, layout_c; std::string to_string() const; SizeArgs(MatrixMulForwardImpl* opr, const TensorLayout& A, const TensorLayout& B, const TensorLayout& C); bool can_be_treated_as_int8x8x32() const { return layout_a.dtype.enumv() == layout_b.dtype.enumv() && (layout_a.dtype.enumv() == DTypeEnum::Int8 || layout_a.dtype.enumv() == DTypeEnum::QuantizedS8) && (layout_c.dtype.enumv() == DTypeEnum::Int32 || layout_c.dtype.enumv() == DTypeEnum::QuantizedS32) && opr->param().format == param::MatrixMul::Format::DEFAULT; } }; struct ExecArgs : public SizeArgs { TensorND tensor_a, tensor_b, tensor_c; Workspace workspace; ExecArgs(MatrixMulForwardImpl* opr, _megdnn_tensor_in A, _megdnn_tensor_in B, _megdnn_tensor_out C, _megdnn_workspace workspace); }; virtual bool is_available(const SizeArgs& args) const = 0; virtual size_t get_workspace_in_bytes(const SizeArgs& args) const = 0; virtual void exec(const ExecArgs& args) const = 0; bool is_available_wk(const SizeArgs& args, size_t limit) { return is_available(args) && get_workspace_in_bytes(args) <= limit; } bool is_available_reproducible( const SizeArgs& args, bool reproducible = true, size_t limit = std::numeric_limits::max()) { return (!reproducible || is_reproducible()) && is_available_wk(args, limit); } AlgoBase& check_workspace(const SizeArgs& args, const Workspace& workspace) { auto req = get_workspace_in_bytes(args); megdnn_assert( req <= workspace.size, "matrix mul fwd algo %s: required workspace %zu bytes, got %zu", name(), req, workspace.size); return *this; } }; class MatrixMulForwardImpl::AlgoCuBlas final : public AlgoBase { public: AlgoCuBlas() = default; bool is_available(const SizeArgs& args) const override; size_t get_workspace_in_bytes(const SizeArgs& /* args */) const override { return 0_z; } const char* name() const override { return "CUBLAS"; } void exec(const ExecArgs& args) const override; bool is_reproducible() const override { return true; } }; #if CUDA_VERSION >= 10000 class MatrixMulForwardImpl::AlgoUInt4x4x32WMMA final : public AlgoBase { public: AlgoUInt4x4x32WMMA() = default; bool is_available(const SizeArgs& args) const override; size_t get_workspace_in_bytes(const SizeArgs& args) const override; const char* name() const override { return "UINT4x4x32_WMMA"; } void exec(const ExecArgs& args) const override; bool is_reproducible() const override { return true; } }; #endif #if CUDA_VERSION >= 10010 class MatrixMulForwardImpl::AlgoCuBlasLt final : public AlgoBase { public: bool is_available(const SizeArgs& args) const override; size_t get_workspace_in_bytes(const SizeArgs& args) const override; const char* name() const override { return "CUBLAS_LT"; } void exec(const ExecArgs& args) const override; bool is_reproducible() const override { return true; } }; #endif class MatrixMulForwardImpl::AlgoNaive final : public AlgoBase { public: AlgoNaive() = default; bool is_available(const SizeArgs& args) const override; size_t get_workspace_in_bytes(const SizeArgs& /* args */) const override { return 0_z; } const char* name() const override { return "NAIVE"; } void exec(const ExecArgs& args) const override; bool is_reproducible() const override { return true; } }; class MatrixMulForwardImpl::AlgoPack { AlgoPack(const AlgoPack&) = delete; AlgoPack& operator=(const AlgoPack&) = delete; public: AlgoPack(); AlgoCuBlas cublas; AlgoNaive naive; #if CUDA_VERSION >= 10000 AlgoUInt4x4x32WMMA wmma_uint4x4x32; #endif #if CUDA_VERSION >= 10010 AlgoCuBlasLt cublas_lt; #endif std::vector all_algos; }; } // namespace cuda } // namespace megdnn // vim: syntax=cpp.doxygen