/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #include "paddle/fluid/operators/benchmark/op_tester.h" #include #include "gflags/gflags.h" #include "gtest/gtest.h" #include "paddle/fluid/framework/op_info.h" #include "paddle/fluid/framework/op_registry.h" #include "paddle/fluid/framework/variable_helper.h" #include "paddle/fluid/platform/init.h" #include "paddle/fluid/platform/profiler.h" #include "paddle/fluid/platform/timer.h" #include "paddle/fluid/pybind/pybind.h" namespace paddle { namespace operators { namespace benchmark { DEFINE_string(op_config_list, "", "Path of op config file."); DEFINE_int32(specified_config_id, -1, "Test the specified op config."); void OpTester::Init(const std::string &filename) { Init(OpTesterConfig(filename)); } void OpTester::Init(const OpTesterConfig &config) { config_ = config; auto &op_desc_info = framework::OpInfoMap::Instance(); // Initialize the OpDesc if (op_desc_info.Has(config_.op_type)) { type_ = config_.op_type; CreateOpDesc(); CreateInputVarDesc(); CreateOutputVarDesc(); } else { LOG(FATAL) << "Op \"" << config_.op_type << "\" is not registered."; } if (config_.device_id >= 0) { place_ = paddle::platform::CUDAPlace(config_.device_id); } else { place_ = paddle::platform::CPUPlace(); } framework::InitDevices(false); scope_.reset(new paddle::framework::Scope()); op_ = framework::OpRegistry::CreateOp(op_desc_); CreateVariables(scope_.get()); } void OpTester::Run() { if (config_.print_debug_string) { LOG(INFO) << DebugString(); } // Warm up RunImpl(); platform::Timer timer; if (config_.profile) { if (platform::is_cpu_place(place_)) { platform::EnableProfiler(platform::ProfilerState::kCPU); } else { #ifdef PADDLE_WITH_CUDA platform::EnableProfiler(platform::ProfilerState::kAll); platform::SetDeviceId(config_.device_id); #else PADDLE_THROW("'CUDAPlace' is not supported in CPU only device."); #endif } timer.Start(); for (int i = config_.repeat; i > 0; --i) { RunImpl(); } timer.Pause(); platform::DisableProfiler(platform::EventSortingKey::kDefault, "op_tester_profiler"); } else { timer.Start(); for (int i = config_.repeat; i > 0; --i) { RunImpl(); } timer.Pause(); } config_.runtime = timer.ElapsedMS() / config_.repeat; LOG(INFO) << "=== Run " << config_.repeat << " times, latency: " << config_.runtime << " ms ==="; } void OpTester::RunImpl() { op_->Run(*scope_, place_); platform::DeviceContextPool::Instance().Get(place_)->Wait(); scope_->DropKids(); } std::vector OpTester::GetOpProtoInputNames() { std::vector input_names; const framework::proto::OpProto &proto = framework::OpInfoMap::Instance().Get(type_).Proto(); for (int i = 0; i != proto.inputs_size(); ++i) { const auto &input = proto.inputs(i); input_names.push_back(input.name()); } return input_names; } std::vector OpTester::GetOpProtoOutputNames() { std::vector output_names; const framework::proto::OpProto &proto = framework::OpInfoMap::Instance().Get(type_).Proto(); for (int i = 0; i != proto.outputs_size(); ++i) { const auto &output = proto.outputs(i); output_names.push_back(output.name()); } return output_names; } std::unordered_map OpTester::GetOpProtoAttrNames() { std::unordered_map attr_types; const framework::proto::OpProto &proto = framework::OpInfoMap::Instance().Get(type_).Proto(); const std::vector skipped_attrs = { framework::OpProtoAndCheckerMaker::OpRoleAttrName(), framework::OpProtoAndCheckerMaker::OpRoleVarAttrName(), framework::OpProtoAndCheckerMaker::OpNamescopeAttrName(), framework::OpProtoAndCheckerMaker::OpCreationCallstackAttrName()}; for (int i = 0; i != proto.attrs_size(); ++i) { const auto &attr = proto.attrs(i); if (!Has(skipped_attrs, attr.name())) { VLOG(4) << "attr: " << attr.name() << ", type: " << attr.type(); attr_types[attr.name()] = attr.type(); } } return attr_types; } framework::proto::VarType::Type OpTester::TransToVarType(std::string str) { if (str == "int32") { return framework::proto::VarType::INT32; } else if (str == "int64") { return framework::proto::VarType::INT64; } else if (str == "fp32") { return framework::proto::VarType::FP32; } else if (str == "fp64") { return framework::proto::VarType::FP64; } else { PADDLE_THROW("Unsupported dtype %s.", str.c_str()); } } void OpTester::CreateInputVarDesc() { std::vector input_names = GetOpProtoInputNames(); for (auto &name : input_names) { const OpInputConfig *input = config_.GetInput(name); if (input == nullptr) { LOG(FATAL) << "The input " << name << " of op " << config_.op_type << " is not correctlly provided."; } std::string var_name = config_.op_type + "." + name; framework::VarDesc *var = Var(var_name); // Need to support more type var->SetType(framework::proto::VarType::LOD_TENSOR); var->SetPersistable(false); var->SetDataType(TransToVarType(input->dtype)); var->SetShape(input->dims); op_desc_.SetInput(name, {var_name}); inputs_[var_name] = *input; } } void OpTester::CreateOutputVarDesc() { std::vector output_names = GetOpProtoOutputNames(); for (auto &name : output_names) { std::string var_name = config_.op_type + "." + name; framework::VarDesc *var = Var(var_name); // Need to support more type var->SetType(framework::proto::VarType::LOD_TENSOR); var->SetPersistable(false); var->SetDataType(framework::proto::VarType::FP32); op_desc_.SetOutput(name, {var_name}); } } void OpTester::CreateOpDesc() { op_desc_.SetType(config_.op_type); std::unordered_map attr_types = GetOpProtoAttrNames(); for (auto item : config_.attrs) { const std::string &name = item.first; if (attr_types.find(name) == attr_types.end()) { LOG(FATAL) << "Operator " << type_ << " do not have attr " << name; } const std::string &value_str = item.second; const framework::proto::AttrType &type = attr_types[name]; switch (type) { case framework::proto::AttrType::BOOLEAN: break; case framework::proto::AttrType::INT: { int value = StringTo(value_str); op_desc_.SetAttr(name, {value}); } break; case framework::proto::AttrType::FLOAT: { float value = StringTo(value_str); op_desc_.SetAttr(name, {value}); } break; case framework::proto::AttrType::STRING: { op_desc_.SetAttr(name, {value_str}); } break; case framework::proto::AttrType::BOOLEANS: case framework::proto::AttrType::INTS: case framework::proto::AttrType::FLOATS: case framework::proto::AttrType::STRINGS: LOG(FATAL) << "Not supported yet."; break; case framework::proto::AttrType::LONG: { int64_t value = StringTo(value_str); op_desc_.SetAttr(name, value); } break; case framework::proto::AttrType::LONGS: default: PADDLE_THROW("Unsupport attr type %d", type); } } } framework::VarDesc *OpTester::Var(const std::string &name) { auto it = vars_.find(name); if (it != vars_.end()) { return it->second.get(); } auto *var = new framework::VarDesc(name); vars_[name].reset(var); return var; } template void OpTester::SetupTensor(framework::LoDTensor *tensor, const std::vector &shape, T lower, T upper, const std::string &initializer, const std::string &filename) { static unsigned int seed = 100; std::mt19937 rng(seed++); std::uniform_real_distribution uniform_dist(0, 1); T *ptr = tensor->mutable_data(framework::make_ddim(shape), place_); framework::LoDTensor cpu_tensor; T *cpu_ptr = nullptr; if (!platform::is_cpu_place(place_)) { cpu_ptr = cpu_tensor.mutable_data(framework::make_ddim(shape), platform::CPUPlace()); } else { cpu_ptr = ptr; } if (initializer == "random") { for (int i = 0; i < cpu_tensor.numel(); ++i) { cpu_ptr[i] = static_cast(uniform_dist(rng) * (upper - lower) + lower); } } else if (initializer == "natural") { for (int i = 0; i < cpu_tensor.numel(); ++i) { cpu_ptr[i] = static_cast(lower + i); } } else if (initializer == "zeros") { for (int i = 0; i < cpu_tensor.numel(); ++i) { cpu_ptr[i] = static_cast(0); } } else if (initializer == "file") { std::ifstream is(filename); for (int i = 0; i < cpu_tensor.numel(); ++i) { T value; is >> value; cpu_ptr[i] = static_cast(value); } is.close(); } else { PADDLE_THROW("Unsupported initializer %s.", initializer.c_str()); } if (!platform::is_cpu_place(place_)) { TensorCopySync(cpu_tensor, place_, tensor); } } void OpTester::CreateVariables(framework::Scope *scope) { for (auto &item : vars_) { auto &var = item.second; if (var->Name() == framework::kEmptyVarName) { continue; } auto *ptr = scope->Var(var->Name()); framework::InitializeVariable(ptr, var->GetType()); if (var->Persistable()) { VLOG(3) << "Create Variable " << var->Name() << " global, which pointer is " << ptr; } else { VLOG(3) << "Create Variable " << var->Name() << " locally, which pointer is " << ptr; } } for (auto &item : inputs_) { // Allocate memory for input tensor auto &var_name = item.first; VLOG(3) << "Allocate memory for tensor " << var_name; auto &var_desc = vars_[var_name]; std::vector shape = var_desc->GetShape(); auto *var = scope->Var(var_name); auto *tensor = var->GetMutable(); const auto &data_type = var_desc->GetDataType(); if (data_type == framework::proto::VarType::INT32) { SetupTensor(tensor, shape, 0, 1, item.second.initializer, item.second.filename); } else if (data_type == framework::proto::VarType::INT64) { SetupTensor(tensor, shape, 0, 1, item.second.initializer, item.second.filename); } else if (data_type == framework::proto::VarType::FP32) { SetupTensor(tensor, shape, static_cast(0.0), static_cast(1.0), item.second.initializer, item.second.filename); } else if (data_type == framework::proto::VarType::FP64) { SetupTensor(tensor, shape, static_cast(0.0), static_cast(1.0), item.second.initializer, item.second.filename); } else { PADDLE_THROW("Unsupported dtype %d.", data_type); } VLOG(3) << "Set lod for tensor " << var_name; std::vector> &lod_vec = item.second.lod; framework::LoD lod; for (size_t i = 0; i < lod_vec.size(); ++i) { lod.push_back(lod_vec[i]); } tensor->set_lod(lod); } } static std::string GenSpaces(int count) { std::stringstream ss; for (int i = 0; i < count; ++i) { ss << " "; } return ss.str(); } std::string OpTester::DebugString() { std::stringstream ss; int count = 0; for (auto &item : vars_) { auto &var = item.second; ss << GenSpaces(count++) << "vars {\n"; ss << GenSpaces(count) << "name: \"" << var->Name() << "\"\n"; ss << GenSpaces(count++) << "type: {\n"; ss << GenSpaces(count) << "type: LOD_TENSOR\n"; ss << GenSpaces(count++) << "lod_tensor {\n"; ss << GenSpaces(count++) << "tensor {\n"; const auto &data_type = var->GetDataType(); if (data_type == framework::proto::VarType::INT32) { ss << GenSpaces(count) << "data_type: INT32\n"; } else if (data_type == framework::proto::VarType::INT64) { ss << GenSpaces(count) << "data_type: INT64\n"; } else if (data_type == framework::proto::VarType::FP32) { ss << GenSpaces(count) << "data_type: FP32\n"; } else if (data_type == framework::proto::VarType::FP64) { ss << GenSpaces(count) << "data_type: FP64\n"; } std::vector shape = var->GetShape(); for (auto d : shape) { ss << GenSpaces(count) << "dims: " << d << "\n"; } ss << GenSpaces(--count) << "}\n"; ss << GenSpaces(--count) << "}\n"; ss << GenSpaces(--count) << "}\n"; ss << GenSpaces(count) << "persistable: " << var->Persistable() << "\n"; ss << GenSpaces(--count) << "}\n"; } ss << GenSpaces(count++) << "ops {\n"; for (auto &name : op_desc_.InputNames()) { ss << GenSpaces(count++) << "inputs {\n"; ss << GenSpaces(count) << "parameters: \"" << name << "\"\n"; ss << GenSpaces(count) << "arguments: \"" << op_desc_.Input(name)[0] << "\"\n"; ss << GenSpaces(--count) << "}\n"; } for (auto &name : op_desc_.OutputNames()) { ss << GenSpaces(count++) << "outputs {\n"; ss << GenSpaces(count) << "parameters: \"" << name << "\"\n"; ss << GenSpaces(count) << "arguments: \"" << op_desc_.Output(name)[0] << "\"\n"; ss << GenSpaces(--count) << "}\n"; } ss << GenSpaces(count) << "type: " << op_desc_.Type() << "\n"; for (auto &name : op_desc_.AttrNames()) { ss << GenSpaces(count++) << "attrs {\n"; const auto &attr_type = op_desc_.GetAttrType(name); const auto &attr = op_desc_.GetAttr(name); ss << GenSpaces(count) << "name: \"" << name << "\"\n"; switch (attr_type) { case framework::proto::AttrType::BOOLEAN: { ss << GenSpaces(count) << "type: BOOLEAN\n"; ss << GenSpaces(count) << "b: " << boost::get(attr) << "\n"; } break; case framework::proto::AttrType::INT: { ss << GenSpaces(count) << "type: INT\n"; ss << GenSpaces(count) << "i: " << boost::get(attr) << "\n"; } break; case framework::proto::AttrType::FLOAT: { ss << GenSpaces(count) << "type: FLOAT\n"; ss << GenSpaces(count) << "f: " << boost::get(attr) << "\n"; } break; case framework::proto::AttrType::STRING: { ss << GenSpaces(count) << "type: STRING\n"; ss << GenSpaces(count) << "s: \"" << boost::get(attr) << "\"\n"; } break; case framework::proto::AttrType::BOOLEANS: { ss << GenSpaces(count) << "type: BOOLEANS\n"; ss << GenSpaces(count) << "bools: " << "\n"; } break; case framework::proto::AttrType::INTS: { ss << GenSpaces(count) << "type: INTS\n"; ss << GenSpaces(count) << "ints: " << "\n"; } break; case framework::proto::AttrType::FLOATS: { ss << GenSpaces(count) << "type: FLOATS\n"; ss << GenSpaces(count) << "floats: " << "\n"; } break; case framework::proto::AttrType::STRINGS: { ss << GenSpaces(count) << "type: STRINGS\n"; ss << GenSpaces(count) << "strings: " << "\n"; } break; case framework::proto::AttrType::LONG: { ss << GenSpaces(count) << "type: LONG\n"; ss << GenSpaces(count) << "l: " << boost::get(attr) << "\n"; } break; case framework::proto::AttrType::LONGS: { ss << GenSpaces(count) << "type: LONGS\n"; ss << GenSpaces(count) << "longs: " << "\n"; } break; default: PADDLE_THROW("Unsupport attr type %d", attr_type); } ss << GenSpaces(--count) << "}\n"; } ss << GenSpaces(--count) << "}\n"; return ss.str(); } TEST(op_tester, base) { if (!FLAGS_op_config_list.empty()) { std::ifstream fin(FLAGS_op_config_list, std::ios::in | std::ios::binary); PADDLE_ENFORCE(static_cast(fin), "Cannot open file %s", FLAGS_op_config_list.c_str()); std::vector op_configs; while (!fin.eof()) { VLOG(4) << "Reading config " << op_configs.size() << "..."; OpTesterConfig config; bool result = config.Init(fin); if (result) { op_configs.push_back(config); } } if (FLAGS_specified_config_id >= 0 && FLAGS_specified_config_id < static_cast(op_configs.size())) { OpTester tester; tester.Init(op_configs[FLAGS_specified_config_id]); tester.Run(); } else { for (size_t i = 0; i < op_configs.size(); ++i) { OpTester tester; tester.Init(op_configs[i]); tester.Run(); } } } else { OpTester tester; OpTesterConfig config; config.op_type = "elementwise_add"; config.inputs.resize(2); config.inputs[0].name = "X"; config.inputs[0].dims = {64, 64}; config.inputs[1].name = "Y"; config.inputs[1].dims = {64, 1}; tester.Init(config); tester.Run(); } } } // namespace benchmark } // namespace operators } // namespace paddle