/** * \file dnn/src/cuda/deformable_conv/bwd_flt/algo.h * MegEngine is Licensed under the Apache License, Version 2.0 (the "License") * * Copyright (c) 2014-2021 Megvii Inc. All rights reserved. * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT ARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. */ #pragma once #include "megdnn/oprs.h" #include "src/common/algo_base.h" #include "src/common/metahelper.h" #include "src/common/utils.h" #include "src/cuda/handle.h" #include "src/cuda/deformable_conv/opr_impl.h" #include namespace megdnn { namespace cuda { class DeformableConvBackwardFilterImpl::AlgoBase : public Algorithm { protected: ~AlgoBase() = default; public: enum class AlgoType : uint32_t { CUDA_MATMUL, }; using Mapper = std::unordered_map; AlgoBase() : Algorithm() { m_handle_type = Handle::HandleType::CUDA; } struct SizeArgs { DeformableConvBackwardFilterImpl* opr; HandleImpl* handle; const TensorLayout& im_layout; const TensorLayout& offset_layout; const TensorLayout& mask_layout; const TensorLayout& out_grad_layout; CanonizedFilterMeta filter_grad_meta; std::string to_string() const; SizeArgs(DeformableConvBackwardFilterImpl* opr, const TensorLayout& im, const TensorLayout& offset, const TensorLayout& mask, const TensorLayout& out_grad, const TensorLayout& filter_grad); SizeArgs(DeformableConvBackwardFilterImpl* opr, const TensorLayout& im, const TensorLayout& offset, const TensorLayout& mask, const TensorLayout& out_grad, const CanonizedFilterMeta& filter_grad_meta); }; struct ExecArgs : public SizeArgs { const TensorND im_tensor, offset_tensor, mask_tensor, out_grad_tensor; TensorND filter_grad_tensor; Workspace workspace; ExecArgs(DeformableConvBackwardFilterImpl* opr, _megdnn_tensor_in im, _megdnn_tensor_in offset, _megdnn_tensor_in mask, _megdnn_tensor_in out_grad, _megdnn_tensor_out filter_grad, _megdnn_workspace workspace); }; virtual bool is_available(const SizeArgs& args) const = 0; virtual size_t get_workspace_in_bytes(const SizeArgs& args) const = 0; virtual void exec(const ExecArgs& args) const = 0; bool is_available_wk(const SizeArgs& args, size_t limit) { return is_available(args) && get_workspace_in_bytes(args) <= limit; } bool is_available_attribute( const SizeArgs& args, const AlgoAttribute& attr = AlgoAttribute::REPRODUCIBLE, size_t limit = std::numeric_limits::max()) { return contain_attribute(attr) && is_available_wk(args, limit); } AlgoBase& check_workspace(const SizeArgs& args, const Workspace& workspace) { auto req = get_workspace_in_bytes(args); megdnn_assert(req <= workspace.size, "deformable_conv bwd_flt algo %s: required workspace %zu " "bytes, got %zu", name(), req, workspace.size); return *this; } }; class DeformableConvBackwardFilterImpl::AlgoMatmul final : public AlgoBase { private: static WorkspaceBundle get_bundle(const SizeArgs& args); public: bool is_available(const SizeArgs& args) const override; size_t get_workspace_in_bytes(const SizeArgs& args) const override; void exec(const ExecArgs& args) const override; AlgoAttribute attribute() const override { return AlgoAttribute::REPRODUCIBLE; } std::vector get_subopr_list( const TensorLayoutArray& layouts, const OperatorBase* opr) const override; const char* name() const override { return "MATMUL"; } MEGDNN_DECL_ALGO_TYPE(CUDA_MATMUL) }; class DeformableConvBackwardFilterImpl::AlgoPack : NonCopyableObj { AlgoBase::Mapper m_all_algos_map; public: AlgoPack(); AlgoMatmul algo_matmul; //! all algorithms std::vector all_algos; const AlgoBase::Mapper& all_algos_map() const { return m_all_algos_map; } }; } // namespace cuda } // namespace megdnn // vim: syntax=cpp.doxygen