/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include "gflags/gflags.h" #include "glog/logging.h" #include "paddle/fluid/framework/tensor.h" #include "paddle/fluid/operators/jit/kernels.h" #include "paddle/fluid/platform/device_tracer.h" #include "paddle/fluid/platform/place.h" #include "paddle/fluid/platform/port.h" #include "paddle/fluid/platform/variant.h" // for UNUSED DEFINE_int32(burning, 10, "Burning times."); DEFINE_int32(repeat, 3000, "Repeat times."); DEFINE_int32(max_size, 1000, "The Max size would be tested."); DEFINE_string(filter, "", "The Benchmark name would be run."); class BenchJITKernel { public: BenchJITKernel() = default; virtual ~BenchJITKernel() = default; virtual void Run() = 0; virtual const char* Name() = 0; virtual const char* Dtype() = 0; virtual const char* Place() = 0; }; static std::vector g_all_benchmarks; BenchJITKernel* InsertBenchmark(BenchJITKernel* b) { g_all_benchmarks.push_back(b); return b; } #define BENCH_JITKERNEL(name, dtype, place) \ class BenchJITKernel_##name##_##dtype##_##place##_ : public BenchJITKernel { \ public: \ const char* Name() override { return #name; } \ const char* Dtype() override { return #dtype; } \ const char* Place() override { return #place; } \ void Run() override; \ }; \ static auto inserted_##name##_##dtype##_##place##_ UNUSED = \ InsertBenchmark(new BenchJITKernel_##name##_##dtype##_##place##_()); \ void BenchJITKernel_##name##_##dtype##_##place##_::Run() #define BENCH_FP32_CPU(name) BENCH_JITKERNEL(name, FP32, CPU) void RUN_ALL_BENCHMARK() { for (auto p : g_all_benchmarks) { if (!FLAGS_filter.empty() && FLAGS_filter != p->Name()) { continue; } LOG(INFO) << "Benchmark " << p->Name() << "." << p->Dtype() << "." << p->Place(); p->Run(); } } template void RandomVec(const int n, T* a, const T lower = static_cast(-20.f), const T upper = static_cast(20.f), unsigned int seed = 100) { std::mt19937 rng(seed); std::uniform_real_distribution uniform_dist(0, 1); for (int i = 0; i < n; ++i) { a[i] = static_cast(uniform_dist(rng) * (upper - lower) + lower); } } std::vector TestSizes() { std::vector s; for (int i = 1; i <= FLAGS_max_size; ++i) { s.push_back(i); } return s; } template struct BenchFunc { // return this function avg time // TODO(TJ): clear cache every time double operator()(const typename KernelTuples::func_type tgt, Args... args) { for (int i = 0; i < FLAGS_burning; ++i) { tgt(args...); } auto start = paddle::platform::PosixInNsec() * 1e-3; for (int i = 0; i < FLAGS_repeat; ++i) { tgt(args...); } auto end = paddle::platform::PosixInNsec() * 1e-3; return static_cast(end - start) / FLAGS_repeat; } }; namespace jit = paddle::operators::jit; template void BenchAllImpls(const typename KernelTuples::attr_type& attr, Args... args) { BenchFunc benchmark; std::vector> infos; // test refer auto refer = jit::GetRefer(); if (!refer) { LOG(FATAL) << "Refer can not be empty!"; } infos.push_back(std::make_pair("Refer", benchmark(refer, args...))); // test jitcode auto jitcode = jit::GetJitCode(attr); if (jitcode) { infos.push_back(std::make_pair("JitCode", benchmark(jitcode, args...))); } // test all impls in more jit::KernelKey kkey(KT, PlaceType()); auto& pool = jit::KernelPool().Instance().AllKernels(); auto iter = pool.find(kkey); if (iter != pool.end()) { auto& impls = iter->second; for (auto& impl : impls) { auto i = dynamic_cast*>(impl.get()); if (i && i->UseMe(attr)) { auto more = i->GetFunc(); infos.push_back( std::make_pair(i->ImplType(), benchmark(more, args...))); } } } // Test result from Get function auto tgt = jit::Get(attr); if (!tgt) { LOG(FATAL) << "Target can not be empty!"; } infos.push_back(std::make_pair("Target", benchmark(tgt, args...))); // print std::ostringstream loginfos; loginfos << "Kernel Type " << jit::to_string(KT) << ": " << attr << ": "; for (auto pair : infos) { loginfos << pair.first << " takes " << pair.second << " us; "; } LOG(INFO) << loginfos.str(); } using Tensor = paddle::framework::Tensor; template void BenchXYZNKernel() { for (int d : TestSizes()) { Tensor x, y, z; x.Resize({d}); y.Resize({d}); z.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); T* z_data = z.mutable_data(PlaceType()); RandomVec(d, x_data); RandomVec(d, y_data); BenchAllImpls, PlaceType>(d, x.data(), y.data(), z_data, d); // test inplace BenchAllImpls, PlaceType>(d, x.data(), z_data, z_data, d); } } template void BenchAXYNKernel() { for (int d : TestSizes()) { const T a = static_cast(3); Tensor x, y; x.Resize({d}); y.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); RandomVec(d, x_data); BenchAllImpls, PlaceType>(d, &a, x.data(), y_data, d); // test inplace BenchAllImpls, PlaceType>(d, &a, x.data(), x_data, d); } } template void BenchXRNKernel() { for (int d : TestSizes()) { Tensor x; RandomVec(d, x.mutable_data({d}, PlaceType())); T res; BenchAllImpls, PlaceType>(d, x.data(), &res, d); } } template void BenchXYNKernel() { for (int d : TestSizes()) { Tensor x, y; x.Resize({d}); y.Resize({d}); T* x_data = x.mutable_data(PlaceType()); T* y_data = y.mutable_data(PlaceType()); RandomVec(d, x_data); BenchAllImpls, PlaceType>(d, x.data(), y_data, d); } } template void BenchLSTMKernel() { for (bool use_peephole : {true, false}) { for (int d : TestSizes()) { const jit::lstm_attr_t attr(d, jit::kVSigmoid, jit::kVTanh, jit::kVTanh, use_peephole); Tensor x, ct_1, ct, ht, wp, checked; x.Resize({4 * d}); ct_1.Resize({d}); ct.Resize({d}); ht.Resize({d}); wp.Resize({3 * d}); checked.Resize({2 * d}); auto place = PlaceType(); RandomVec(x.numel(), x.mutable_data(place), -2.f, 2.f); RandomVec(wp.numel(), wp.mutable_data(place), -2.f, 2.f); RandomVec(ct_1.numel(), ct_1.mutable_data(place), -2.f, 2.f); const T* ct_1_data = ct_1.data(); const T* wp_data = wp.data(); T* x_data = x.mutable_data(place); T* checked_data = checked.mutable_data(place); T* ct_data = ct.mutable_data(place); T* ht_data = ht.mutable_data(place); jit::lstm_t step; step.gates = x_data; step.ct_1 = ct_1_data; step.ct = ct_data; step.ht = ht_data; if (use_peephole) { step.wp = wp_data; step.checked = checked_data; } BenchAllImpls, PlaceType>(attr, &step, &attr); } } } template void BenchGRUKernel() { for (int d : TestSizes()) { const jit::gru_attr_t attr(d, jit::kVSigmoid, jit::kVTanh); auto place = PlaceType(); Tensor x, ht_1, ht; x.Resize({3 * d}); ht_1.Resize({d}); ht.Resize({d}); RandomVec(3 * d, x.mutable_data(place), -2.f, 2.f); RandomVec(d, ht_1.mutable_data(place), -2.f, 2.f); const T* ht_1_data = ht_1.data(); T* x_data = x.mutable_data(place); T* ht_data = ht.mutable_data(place); jit::gru_t step; step.gates = x_data; step.ht_1 = ht_1_data; step.ht = ht_data; BenchAllImpls, PlaceType>(attr, &step, &attr); } } template void BenchSeqPoolKernel() { std::vector pool_types = { jit::SeqPoolType::kSum, jit::SeqPoolType::kAvg, jit::SeqPoolType::kSqrt}; for (auto type : pool_types) { for (int w : TestSizes()) { jit::seq_pool_attr_t attr(w, type); for (int h : TestSizes()) { attr.h = h; Tensor x, y; x.Resize({h * w}); y.Resize({w}); RandomVec(h * w, x.mutable_data(PlaceType()), -2.f, 2.f); const T* x_data = x.data(); T* y_data = y.mutable_data(PlaceType()); BenchAllImpls, PlaceType>(attr, x_data, y_data, &attr); } } } } template void BenchEmbSeqPoolKernel() { std::vector pool_types = {jit::SeqPoolType::kSum}; int64_t tbl_h = 1e4; for (int tbl_w : {10, 16, 256}) { Tensor table; table.Resize({tbl_h, tbl_w}); RandomVec(tbl_h * tbl_w, table.mutable_data(PlaceType()), -2.f, 2.f); const T* table_data = table.data(); for (auto type : pool_types) { for (int idx_w : {1, 2, 10, 16}) { for (int idx_h : {1, 2, 9, 13, 16}) { int64_t out_w = tbl_w * idx_w; jit::emb_seq_pool_attr_t attr(tbl_h, tbl_w, idx_h, idx_w, out_w, type); Tensor idx, out; idx.Resize({idx_h, idx_w}); out.Resize({out_w}); RandomVec(idx_h * idx_w, idx.mutable_data(PlaceType()), 0, tbl_h - 1); const int64_t* idx_data = idx.data(); T* o_data = out.mutable_data(PlaceType()); BenchAllImpls, PlaceType>( attr, table_data, idx_data, o_data, &attr); } } } } } template void BenchSgdKernel() { const T lr = 0.1; auto UnDuplicatedRandomVec = [](int n, const int64_t lower, const int64_t upper) -> std::vector { PADDLE_ENFORCE_LE(static_cast(upper - lower), n - 1); PADDLE_ENFORCE_GT(n, 0); std::vector all, out; for (int i = 0; i < n; ++i) { all.push_back(i); } std::random_shuffle(all.begin(), all.end()); out.insert(out.begin(), all.begin(), all.begin() + n); return out; }; for (int param_h : {1, 1000}) { for (int grad_w : {1, 2, 8, 16, 30, 256}) { // only benchmark inplace Tensor param; param.Resize({param_h, grad_w}); T* param_data = param.mutable_data(PlaceType()); RandomVec(param_h * grad_w, param_data, -2.f, 2.f); for (int rows_size = 1; rows_size <= std::min(param_h, 10); ++rows_size) { Tensor grad; grad.Resize({rows_size, grad_w}); std::vector rows = UnDuplicatedRandomVec(rows_size, 0, rows_size - 1); RandomVec(rows_size * grad_w, grad.mutable_data(PlaceType()), -2.f, 2.f); const T* grad_data = grad.data(); const int64_t* rows_data = rows.data(); jit::sgd_attr_t attr(param_h, grad_w, rows_size, grad_w, rows_size); BenchAllImpls, PlaceType>( attr, &lr, param_data, grad_data, rows_data, param_data, &attr); } } } } template void BenchMatMulKernel() { for (int m : {1, 2, 3, 4}) { for (int n : TestSizes()) { for (int k : TestSizes()) { Tensor a, b, c; a.Resize({m * k}); b.Resize({k * n}); c.Resize({m * n}); RandomVec(m * k, a.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(k * n, b.mutable_data(PlaceType()), -2.f, 2.f); const T* a_data = a.data(); const T* b_data = b.data(); T* c_data = c.mutable_data(PlaceType()); const jit::matmul_attr_t attr{m, n, k}; BenchAllImpls, PlaceType>(attr, a_data, b_data, c_data, &attr); } } } } template void BenchSoftmaxKernel() { for (int bs : {1, 2, 10}) { for (int n : TestSizes()) { Tensor x, y; x.Resize({bs, n}); y.Resize({bs, n}); RandomVec(bs * n, x.mutable_data(PlaceType()), -2.f, 2.f); const T* x_data = x.data(); T* y_data = y.mutable_data(PlaceType()); BenchAllImpls, PlaceType>(n, x_data, y_data, n, bs); } } } template void BenchLayerNormKernel() { const T epsilon = 9.99999975e-06; for (int n : {1, 2, 10}) { for (int x_dim_0 : {1, 9, 17, 50}) { int left = n * x_dim_0; for (int x_dim_1 : TestSizes()) { int right = x_dim_1; int sz = left * right; Tensor x, mean, var, scale, bias, out; x.Resize({n, x_dim_0, x_dim_1}); out.Resize({n, x_dim_0, x_dim_1}); mean.Resize({n, x_dim_0}); var.Resize({n, x_dim_0}); scale.Resize({x_dim_1}); bias.Resize({x_dim_1}); RandomVec(sz, x.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(left, mean.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(left, var.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(right, scale.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(right, bias.mutable_data(PlaceType()), -2.f, 2.f); const T* scale_data = scale.data(); const T* bias_data = bias.data(); T* x_data = x.data(); T* mean_data = mean.data(); T* var_data = var.data(); T* out_data = out.mutable_data(PlaceType()); BenchAllImpls, PlaceType>( right, x_data, out_data, mean_data, var_data, scale_data, bias_data, left, epsilon, right); } } } } template void BenchCRFDecodingKernel() { constexpr int state_trans_base_idx = 2; for (int seq_len : {1, 11, 17, 50}) { for (int tag_num : TestSizes()) { int x_sz = seq_len * tag_num; int w_sz = (tag_num + state_trans_base_idx) * tag_num; Tensor x, w, alpha, track; x.Resize({seq_len, tag_num}); w.Resize({tag_num + state_trans_base_idx, tag_num}); alpha.Resize({seq_len, tag_num}); track.Resize({seq_len, tag_num}); RandomVec(x_sz, x.mutable_data(PlaceType()), -2.f, 2.f); RandomVec(w_sz, w.mutable_data(PlaceType()), -2.f, 2.f); const T* x_data = x.data(); const T* w_data = w.data(); T* alpha_data = alpha.mutable_data(PlaceType()); int* track_data = track.mutable_data(PlaceType()); BenchAllImpls, PlaceType>( tag_num, seq_len, x_data, w_data, alpha_data, track_data, tag_num); } } } using T = float; using CPUPlace = paddle::platform::CPUPlace; // xyzn BENCH_FP32_CPU(kVMul) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVAdd) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVAddRelu) { BenchXYZNKernel(); } BENCH_FP32_CPU(kVSub) { BenchXYZNKernel(); } // axyn BENCH_FP32_CPU(kVScal) { BenchAXYNKernel(); } BENCH_FP32_CPU(kVAddBias) { BenchAXYNKernel(); } // xrn BENCH_FP32_CPU(kHSum) { BenchXRNKernel(); } BENCH_FP32_CPU(kHMax) { BenchXRNKernel(); } // xyn BENCH_FP32_CPU(kVRelu) { BenchXYNKernel(); } BENCH_FP32_CPU(kVIdentity) { BenchXYNKernel(); } BENCH_FP32_CPU(kVSquare) { BenchXYNKernel(); } BENCH_FP32_CPU(kVExp) { BenchXYNKernel(); } BENCH_FP32_CPU(kVSigmoid) { BenchXYNKernel(); } BENCH_FP32_CPU(kVTanh) { BenchXYNKernel(); } // lstm and peephole BENCH_FP32_CPU(kLSTMCtHt) { BenchLSTMKernel(); } BENCH_FP32_CPU(kLSTMC1H1) { BenchLSTMKernel(); } // gru functions BENCH_FP32_CPU(kGRUH1) { BenchGRUKernel(); } BENCH_FP32_CPU(kGRUHtPart1) { BenchGRUKernel(); } BENCH_FP32_CPU(kGRUHtPart2) { BenchGRUKernel(); } // seq pool function BENCH_FP32_CPU(kSeqPool) { BenchSeqPoolKernel(); } // embedding seq pool function BENCH_FP32_CPU(kEmbSeqPool) { BenchEmbSeqPoolKernel(); } // sgd function BENCH_FP32_CPU(kSgd) { BenchSgdKernel(); } // matmul BENCH_FP32_CPU(kMatMul) { BenchMatMulKernel(); } // softmax BENCH_FP32_CPU(kSoftmax) { BenchSoftmaxKernel(); } // layernorm BENCH_FP32_CPU(kLayerNorm) { BenchLayerNormKernel(); } // crfdecoding BENCH_FP32_CPU(kCRFDecoding) { BenchCRFDecodingKernel(); } // Benchmark all jit kernels including jitcode, mkl and refer. // To use this tool, run command: ./benchmark [options...] // Options: // --burning: the burning time before count // --repeat: the repeat times // --max_size: the max size would be tested // --filter: the bench name would be run int main(int argc, char* argv[]) { gflags::ParseCommandLineFlags(&argc, &argv, true); google::InitGoogleLogging(argv[0]); LOG(INFO) << "Burning " << FLAGS_burning << " times, Repeat " << FLAGS_repeat << " times."; RUN_ALL_BENCHMARK(); }