// Copyright (c) 2022 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include "glog/logging.h" #include "paddle/phi/kernels/autotune/gpu_timer.h" #include "paddle/phi/kernels/autotune/switch_autotune.h" namespace phi { namespace autotune { template class KernelCallback { public: using ReturnT = ReturnType; using FuncType = ReturnType (*)(Args...); KernelCallback() {} explicit KernelCallback(FuncType func_) : func(func_) {} virtual ~KernelCallback() {} ReturnType Run(Args... args) { return func(args...); } private: FuncType func; }; template static KernelCallback MakeCallback( ReturnType (*cb)(Args...)) { return KernelCallback(cb); } template class AutoTuneBase { public: AutoTuneBase() {} virtual ~AutoTuneBase() {} explicit AutoTuneBase(KernelType kernel) { kernels_.push_back(/*default=*/kernel); } template void AddCallBack(ReturnType (*func)(Args...)) { if (!is_init_) { std::lock_guard lock(mutex_); kernels_.push_back(MakeCallback(func)); } } template void Run(const Context& ctx, const AlgorithmType& algo, const size_t key, Args&&... args) { PADDLE_ENFORCE_GT( kernels_.size(), 0, phi::errors::InvalidArgument( "kernel num must be greater than 0, now is %d", kernels_.size())); is_init_ = true; auto& cache = AutoTuneCache::Instance().Get(algo); if (cache.Find(key)) { auto best_idx = cache.Get(key); kernels_[best_idx].Run(args...); } else { bool use_autotune = AutoTuneStatus::Instance().UseAutoTune(); if (use_autotune) { // All avaliable kernels have ran while picking the best kernel, // so there may be no need for another kernel run. auto best_idx = PickBestKernel(ctx, args...); cache.Set(key, best_idx); } else { kernels_[0].Run(args...); } } } private: bool is_init_{false}; std::vector kernels_; mutable std::mutex mutex_; template size_t PickBestKernel(const Context& ctx, Args&&... args) { std::lock_guard lock(mutex_); PADDLE_ENFORCE_GT( kernels_.size(), 0, phi::errors::InvalidArgument( "kernel num must be greater than 0, now is %d", kernels_.size())); size_t best_idx = 0; float min_time = std::numeric_limits::max(); // Time cost test estabulished in default stream. for (int i = 0; i < kernels_.size(); ++i) { auto time = RunAndMeasureKernel(ctx, i, args...); if (time < min_time) { min_time = time; best_idx = i; } } VLOG(3) << "best kernel idx is " << best_idx; return best_idx; } template float RunAndMeasureKernel(const Context& ctx, const int idx, Args&&... args) { // Regard 1st run as warmup, judge the compare result by the time cost // of rest cycles. constexpr int repeats = 6; phi::GpuTimer timer; float time_cost = 0; const auto& stream = ctx.stream(); ctx.Wait(); for (int i = 0; i < repeats; ++i) { timer.Start(stream); kernels_[idx].Run(args...); timer.Stop(stream); auto time = timer.ElapsedTime(); if (i > 0) { time_cost += time; } VLOG(3) << "kernel[" << idx << "][" << i << "th time cost is " << time; } return time_cost; } }; template static AutoTuneBase> MakeAutoTuner( ReturnType (*func)(Args...)) { auto obj = MakeCallback(func); return AutoTuneBase(obj); } template class TransposeAutoTuner : public AutoTuneBase> { public: static AutoTuneBase>* Instance( ReturnType (*func)(Args...)) { static std::once_flag transpose_init_flag_; static std::unique_ptr< AutoTuneBase>> instance_; std::call_once(transpose_init_flag_, [&] { auto obj = MakeCallback(func); instance_.reset(new AutoTuneBase(obj)); }); return instance_.get(); } }; template static AutoTuneBase>* MakeTransposeTuner(ReturnType (*func)(Args...)) { return TransposeAutoTuner::Instance(func); } } // namespace autotune } // namespace phi