From f8d38c14f0f7640574e4bdb048211add6a56872d Mon Sep 17 00:00:00 2001 From: yuanshuai Date: Thu, 27 Jun 2019 11:45:58 +0000 Subject: [PATCH] [LITE][ARM] Add model test bin of arm --- paddle/fluid/lite/api/CMakeLists.txt | 6 ++ paddle/fluid/lite/api/model_test.cc | 143 +++++++++++++++++++++++++++ paddle/fluid/lite/api/test_helper.h | 1 + paddle/fluid/lite/tools/build.sh | 2 +- 4 files changed, 151 insertions(+), 1 deletion(-) create mode 100644 paddle/fluid/lite/api/model_test.cc diff --git a/paddle/fluid/lite/api/CMakeLists.txt b/paddle/fluid/lite/api/CMakeLists.txt index d982f10175..48d8f06529 100644 --- a/paddle/fluid/lite/api/CMakeLists.txt +++ b/paddle/fluid/lite/api/CMakeLists.txt @@ -121,6 +121,12 @@ lite_cc_test(test_paddle_api_lite SRCS paddle_api_test.cc DEPS paddle_api_full p ARM_DEPS ${arm_kernels} X86_DEPS ${x86_kernels} ARGS --model_dir=${LITE_MODEL_DIR}/lite_naive_model SERIAL) + +lite_cc_test(test_model_bin SRCS model_test.cc DEPS paddle_api_full paddle_api_light + ${ops_lite} + ARM_DEPS ${arm_kernels} + X86_DEPS ${x86_kernels}) + if (WITH_TESTING) add_dependencies(test_paddle_api_lite extern_lite_download_lite_naive_model_tar_gz) endif() diff --git a/paddle/fluid/lite/api/model_test.cc b/paddle/fluid/lite/api/model_test.cc new file mode 100644 index 0000000000..78fe52394c --- /dev/null +++ b/paddle/fluid/lite/api/model_test.cc @@ -0,0 +1,143 @@ +// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. +// +// Licensed under the Apache License, Version 2.0 (the "License"); +// you may not use this file except in compliance with the License. +// You may obtain a copy of the License at +// +// http://www.apache.org/licenses/LICENSE-2.0 +// +// Unless required by applicable law or agreed to in writing, software +// distributed under the License is distributed on an "AS IS" BASIS, +// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +// See the License for the specific language governing permissions and +// limitations under the License. + +#include +#include +#include +#include "paddle/fluid/lite/api/paddle_api.h" +#include "paddle/fluid/lite/api/paddle_use_kernels.h" +#include "paddle/fluid/lite/api/paddle_use_ops.h" +#include "paddle/fluid/lite/api/paddle_use_passes.h" +#include "paddle/fluid/lite/api/test_helper.h" +#include "paddle/fluid/lite/core/cpu_info.h" +#include "paddle/fluid/lite/utils/string.h" + +namespace paddle { +namespace lite_api { + +void OutputOptModel(const std::string& load_model_dir, + const std::string& save_optimized_model_dir, + const std::vector& input_shape) { + lite_api::CxxConfig config; + config.set_model_dir(load_model_dir); + config.set_preferred_place(Place{TARGET(kX86), PRECISION(kFloat)}); + config.set_valid_places({ + Place{TARGET(kX86), PRECISION(kFloat)}, + Place{TARGET(kARM), PRECISION(kFloat)}, + }); + auto predictor = lite_api::CreatePaddlePredictor(config); + + auto input_tensor = predictor->GetInput(0); + input_tensor->Resize(input_shape); + auto* data = input_tensor->mutable_data(); + int input_num = 1; + for (int i = 0; i < input_shape.size(); ++i) { + input_num *= input_shape[i]; + } + for (int i = 0; i < input_num; ++i) { + data[i] = i; + } + predictor->Run(); + // delete old optimized model + int ret = system( + paddle::lite::string_format("rm -rf %s", save_optimized_model_dir.c_str()) + .c_str()); + if (ret == 0) { + LOG(INFO) << "delete old optimized model " << save_optimized_model_dir; + } + predictor->SaveOptimizedModel(save_optimized_model_dir); + LOG(INFO) << "Load model from " << load_model_dir; + LOG(INFO) << "Save optimized model to " << save_optimized_model_dir; +} + +#ifdef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK +void Run(const std::vector& input_shape, const std::string& model_dir, + const int repeat, const int thread_num, const int warmup_times = 10) { + lite::DeviceInfo::Init(); + lite::DeviceInfo::Global().SetRunMode(lite::LITE_POWER_HIGH, thread_num); + lite_api::MobileConfig config; + config.set_model_dir(model_dir); + + auto predictor = lite_api::CreatePaddlePredictor(config); + + auto input_tensor = predictor->GetInput(0); + input_tensor->Resize(input_shape); + float* input_data = input_tensor->mutable_data(); + int input_num = 1; + for (int i = 0; i < input_shape.size(); ++i) { + input_num *= input_shape[i]; + } + for (int i = 0; i < input_num; ++i) { + input_data[i] = i; + } + + for (int i = 0; i < warmup_times; ++i) { + predictor->Run(); + } + + auto start = lite::GetCurrentUS(); + for (int i = 0; i < repeat; ++i) { + predictor->Run(); + } + auto end = lite::GetCurrentUS(); + + LOG(INFO) << "================== Speed Report ==================="; + LOG(INFO) << "Model: " << model_dir << ", threads num " << thread_num + << ", warmup: " << warmup_times << ", repeats: " << repeat + << ", spend " << (end - start) / repeat / 1000.0 + << " ms in average."; + + auto output = predictor->GetOutput(0); + const float* out = output->data(); + LOG(INFO) << "out " << out[0]; + LOG(INFO) << "out " << out[1]; + auto output_shape = output->shape(); + int output_num = 1; + for (int i = 0; i < output_shape.size(); ++i) { + output_num *= output_shape[i]; + } + LOG(INFO) << "output_num: " << output_num; +} +#endif + +} // namespace lite_api +} // namespace paddle + +int main(int argc, char** argv) { + if (argc < 4) { + LOG(INFO) << "usage: " << argv[0] << " "; + exit(0); + } + std::string load_model_dir = argv[1]; + std::string save_optimized_model_dir = load_model_dir + "opt2"; + +#ifdef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK + int repeat = std::stoi(argv[2]); + int thread_num = std::stoi(argv[3]); +#endif + + std::vector input_shape{1, 3, 224, 224}; + + // Output optimized model + paddle::lite_api::OutputOptModel(load_model_dir, save_optimized_model_dir, + input_shape); + +#ifdef LITE_WITH_LIGHT_WEIGHT_FRAMEWORK + // Run inference using optimized model + paddle::lite_api::Run(input_shape, save_optimized_model_dir, repeat, + thread_num); +#endif + + return 0; +} diff --git a/paddle/fluid/lite/api/test_helper.h b/paddle/fluid/lite/api/test_helper.h index 4d184eeb16..1a5ab31abd 100644 --- a/paddle/fluid/lite/api/test_helper.h +++ b/paddle/fluid/lite/api/test_helper.h @@ -15,6 +15,7 @@ #pragma once #include +#include #include // for eval diff --git a/paddle/fluid/lite/tools/build.sh b/paddle/fluid/lite/tools/build.sh index 3bb4b219a2..c832c6304b 100755 --- a/paddle/fluid/lite/tools/build.sh +++ b/paddle/fluid/lite/tools/build.sh @@ -295,7 +295,7 @@ function test_arm { echo "android do not need armv7hf" return 0 fi - + echo "test file: ${TESTS_FILE}" for _test in $(cat $TESTS_FILE); do test_arm_android $_test $port -- GitLab