/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include "gflags/gflags.h" #include "glog/logging.h" #include "paddle/fluid/framework/tensor.h" #include "paddle/fluid/operators/jit/kernels.h" #include "paddle/fluid/platform/device_tracer.h" #include "paddle/fluid/platform/place.h" #include "paddle/fluid/platform/port.h" #include "paddle/fluid/platform/variant.h" // for UNUSED DEFINE_int32(burning, 10, "Burning times."); DEFINE_int32(repeat, 3000, "Repeat times."); DEFINE_int32(max_size, 1000, "The Max size would be tested."); DEFINE_string(filter, "", "The Benchmark name would be run."); class BenchJITKernel { public: BenchJITKernel() = default; virtual ~BenchJITKernel() = default; virtual void Run() = 0; virtual const char* Name() = 0; virtual const char* Dtype() = 0; virtual const char* Place() = 0; }; static std::vector g_all_benchmarks; BenchJITKernel* InsertBenchmark(BenchJITKernel* b) { g_all_benchmarks.push_back(b); return b; } #define BENCH_JITKERNEL(name, dtype, place) \ class BenchJITKernel_##name##_##dtype##_##place##_ : public BenchJITKernel { \ public: \ const char* Name() override { return #name; } \ const char* Dtype() override { return #dtype; } \ const char* Place() override { return #place; } \ void Run() override; \ }; \ static auto inserted_##name##_##dtype##_##place##_ UNUSED = \ InsertBenchmark(new BenchJITKernel_##name##_##dtype##_##place##_()); \ void BenchJITKernel_##name##_##dtype##_##place##_::Run() #define BENCH_FP32_CPU(name) BENCH_JITKERNEL(name, FP32, CPU) void RUN_ALL_BENCHMARK() { for (auto p : g_all_benchmarks) { if (!FLAGS_filter.empty() && FLAGS_filter != p->Name()) { continue; } LOG(INFO) << "Benchmark " << p->Name() << "." << p->Dtype() << "." << p->Place(); p->Run(); } } template void RandomVec(const int n, T* a, const T lower = static_cast(-20.f), const T upper = static_cast(20.f), unsigned int seed = 100) { std::mt19937 rng(seed); std::uniform_real_distribution uniform_dist(0, 1); for (int i = 0; i < n; ++i) { a[i] = static_cast(uniform_dist(rng) * (upper - lower) + lower); } } std::vector TestSizes() { std::vector s; for (int i = 1; i <= FLAGS_max_size; ++i) { s.push_back(i); } return s; } template struct BenchFunc { // return this function avg time // TODO(TJ): clear cache every time double operator()(const typename KernelTuples::func_type tgt, Args... args) { for (int i = 0; i < FLAGS_burning; ++i) { tgt(args...); } auto start = paddle::platform::PosixInNsec() * 1e-3; for (int i = 0; i < FLAGS_repeat; ++i) { tgt(args...); } auto end = paddle::platform::PosixInNsec() * 1e-3; return static_cast(end - start) / FLAGS_repeat; } }; namespace jit = paddle::operators::jit; template void BenchAllImpls(const typename KernelTuples::attr_type& attr, Args... args) { BenchFunc benchmark; std::vector> infos; // test refer auto refer = jit::GetRefer(); if (!refer) { LOG(FATAL) << "Refer can not be empty!"; } infos.push_back(std::make_pair("Refer", benchmark(refer, args...))); // test jitcode auto jitcode = jit::GetJitCode(attr); if (jitcode) { infos.push_back(std::make_pair("JitCode", benchmark(jitcode, args...))); } // test all impls in more jit::KernelKey kkey(KT, PlaceType()); auto& pool = jit::KernelPool().Instance().AllKernels(); auto iter = pool.find(kkey); if (iter != pool.end()) { auto& impls = iter->second; for (auto& impl : impls) { auto i = dynamic_cast*>(impl.get()); if (i && i->UseMe(attr)) { auto more = i->GetFunc(); infos.push_back( std::make_pair(i->ImplType(), benchmark(more, args...))); } } } // Test result from Get function auto tgt = jit::Get(attr); if (!tgt) { LOG(FATAL) << "Target can not be empty!"; } infos.push_back(std::make_pair("Target", benchmark(tgt, args...))); // print std::ostringstream loginfos; loginfos << "Kernel Type " << jit::to_string(KT) << ": " << attr << ": "; for (auto pair : infos) { loginfos << pair.first << " takes " << pair.second << " us; "; } LOG(INFO) << loginfos.str(); } using Tensor = paddle::framework::Tensor; template void BenchXYZNKernel() { for (int d : TestSizes()) { Tensor x, y, z; x.Resize({d}); y.Resize({d}); z.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); T* z_data = z.mutable_data(PlaceType()); RandomVec(d, x_data); RandomVec(d, y_data); BenchAllImpls, PlaceType>(d, x.data(), y.data(), z_data, d); // test inplace BenchAllImpls, PlaceType>(d, x.data(), z_data, z_data, d); } } template void BenchAXYNKernel() { for (int d : TestSizes()) { const T a = static_cast(3); Tensor x, y; x.Resize({d}); y.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); RandomVec(d, x_data); BenchAllImpls, PlaceType>(d, &a, x.data(), y_data, d); // test inplace BenchAllImpls, PlaceType>(d, &a, x.data(), x_data, d); } } template void BenchXRNKernel() { for (int d : TestSizes()) { Tensor x; RandomVec(d, x.mutable_data({d}, PlaceType())); T res; BenchAllImpls, PlaceType>(d, x.data(), &res, d); } } template void BenchXYNKernel() { for (int d : TestSizes()) { Tensor x, y; x.Resize({d}); y.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); RandomVec(d, x_data); BenchAllImpls, PlaceType>(d, x.data(), y_data, d); } } template void BenchLSTMKernel() { for (bool use_peephole : {true, false}) { for (int d : TestSizes()) { const jit::lstm_attr_t attr(d, jit::kVSigmoid, jit::kVTanh, jit::kVTanh, use_peephole); Tensor x, ct_1, ct, ht, wp, checked; x.Resize({4 * d}); ct_1.Resize({d}); ct.Resize({d}); ht.Resize({d}); wp.Resize({3 * d}); checked.Resize({2 * d}); auto place = PlaceType(); RandomVec(x.numel(), x.mutable_data(place), -2.f, 2.f); RandomVec(wp.numel(), wp.mutable_data(place), -2.f, 2.f); RandomVec(ct_1.numel(), ct_1.mutable_data(place), -2.f, 2.f); const T* ct_1_data = ct_1.data(); const T* wp_data = wp.data(); T* x_data = x.mutable_data(place); T* checked_data = checked.mutable_data(place); T* ct_data = ct.mutable_data(place); T* ht_data = ht.mutable_data(place); jit::lstm_t step; step.gates = x_data; step.ct_1 = ct_1_data; step.ct = ct_data; step.ht = ht_data; if (use_peephole) { step.wp = wp_data; step.checked = checked_data; } BenchAllImpls, PlaceType>(attr, &step, &attr); } } } template void BenchGRUKernel() { for (int d : TestSizes()) { const jit::gru_attr_t attr(d, jit::kVSigmoid, jit::kVTanh); auto place = PlaceType(); Tensor x, ht_1, ht; x.Resize({3 * d}); ht_1.Resize({d}); ht.Resize({d}); RandomVec(3 * d, x.mutable_data(place), -2.f, 2.f); RandomVec(d, ht_1.mutable_data(place), -2.f, 2.f); const T* ht_1_data = ht_1.data(); T* x_data = x.mutable_data(place); T* ht_data = ht.mutable_data(place); jit::gru_t step; step.gates = x_data; step.ht_1 = ht_1_data; step.ht = ht_data; BenchAllImpls, PlaceType>(attr, &step, &attr); } } template void BenchSeqPoolKernel() { std::vector pool_types = { jit::SeqPoolType::kSum, jit::SeqPoolType::kAvg, jit::SeqPoolType::kSqrt}; for (auto type : pool_types) { for (int w : TestSizes()) { jit::seq_pool_attr_t attr(w, type); for (int h : TestSizes()) { attr.h = h; Tensor x, y; x.Resize({h * w}); y.Resize({w}); RandomVec(h * w, x.mutable_data(PlaceType()), -2.f, 2.f); const T* x_data = x.data(); T* y_data = y.mutable_data(PlaceType()); BenchAllImpls, PlaceType>(attr, x_data, y_data, &attr); } } } } template void BenchMatMulKernel() { for (int m : {1, 2, 3, 4}) { for (int n : TestSizes()) { for (int k : TestSizes()) { Tensor a, b, c; a.Resize({m * k}); b.Resize({k * n}); c.Resize({m * n}); RandomVec(m * k, a.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(k * n, b.mutable_data(PlaceType()), -2.f, 2.f); const T* a_data = a.data(); const T* b_data = b.data(); T* c_data = c.mutable_data(PlaceType()); BenchAllImpls, PlaceType>(k, a_data, b_data, c_data, m, n, k); } } } } template void BenchSoftmaxKernel() { for (int bs : {1, 2, 10}) { for (int n : TestSizes()) { Tensor x, y; x.Resize({bs, n}); y.Resize({bs, n}); RandomVec(bs * n, x.mutable_data(PlaceType()), -2.f, 2.f); const T* x_data = x.data(); T* y_data = y.mutable_data(PlaceType()); BenchAllImpls, PlaceType>(n, x_data, y_data, n, bs); } } } using T = float; using CPUPlace = paddle::platform::CPUPlace; // xyzn BENCH_FP32_CPU(kVMul) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVAdd) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVAddRelu) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVSub) { BenchXYZNKernel(); } // axyn BENCH_FP32_CPU(kVScal) { BenchAXYNKernel(); } BENCH_FP32_CPU(kVAddBias) { BenchAXYNKernel(); } // xrn BENCH_FP32_CPU(kHSum) { BenchXRNKernel(); } BENCH_FP32_CPU(kHMax) { BenchXRNKernel(); } // xyn BENCH_FP32_CPU(kVRelu) { BenchXYNKernel(); } BENCH_FP32_CPU(kVIdentity) { BenchXYNKernel(); } BENCH_FP32_CPU(kVSquare) { BenchXYNKernel(); } BENCH_FP32_CPU(kVExp) { BenchXYNKernel(); } BENCH_FP32_CPU(kVSigmoid) { BenchXYNKernel(); } BENCH_FP32_CPU(kVTanh) { BenchXYNKernel(); } // lstm and peephole BENCH_FP32_CPU(kLSTMCtHt) { BenchLSTMKernel(); } BENCH_FP32_CPU(kLSTMC1H1) { BenchLSTMKernel(); } // gru functions BENCH_FP32_CPU(kGRUH1) { BenchGRUKernel(); } BENCH_FP32_CPU(kGRUHtPart1) { BenchGRUKernel(); } BENCH_FP32_CPU(kGRUHtPart2) { BenchGRUKernel(); } // seq pool function BENCH_FP32_CPU(kSeqPool) { BenchSeqPoolKernel(); } // matmul BENCH_FP32_CPU(kMatMul) { BenchMatMulKernel(); } // softmax BENCH_FP32_CPU(kSoftmax) { BenchSoftmaxKernel(); } // Benchmark all jit kernels including jitcode, mkl and refer. // To use this tool, run command: ./benchmark [options...] // Options: // --burning: the burning time before count // --repeat: the repeat times // --max_size: the max size would be tested // --filter: the bench name would be run int main(int argc, char* argv[]) { gflags::ParseCommandLineFlags(&argc, &argv, true); google::InitGoogleLogging(argv[0]); LOG(INFO) << "Burning " << FLAGS_burning << " times, Repeat " << FLAGS_repeat << " times."; RUN_ALL_BENCHMARK(); }