From 7eebb905235c5780350d92902776a4e0c267c87f Mon Sep 17 00:00:00 2001 From: tensor-tang Date: Fri, 7 Sep 2018 11:53:19 +0800 Subject: [PATCH] fix conflicts --- paddle/fluid/inference/analysis/analyzer_lac_tester.cc | 2 +- paddle/fluid/inference/api/helper.h | 2 ++ 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/paddle/fluid/inference/analysis/analyzer_lac_tester.cc b/paddle/fluid/inference/analysis/analyzer_lac_tester.cc index 5740faa746..7917152428 100644 --- a/paddle/fluid/inference/analysis/analyzer_lac_tester.cc +++ b/paddle/fluid/inference/analysis/analyzer_lac_tester.cc @@ -192,7 +192,7 @@ void TestLACPrediction(const std::string &model_path, sum += timer.toc(); } } - PrintTime(sum, batch_size, repeat); + PrintTime(batch_size, repeat, 1, 0, sum / batch_size); return; } timer.tic(); diff --git a/paddle/fluid/inference/api/helper.h b/paddle/fluid/inference/api/helper.h index 2c2ac656e8..0ab2542f34 100644 --- a/paddle/fluid/inference/api/helper.h +++ b/paddle/fluid/inference/api/helper.h @@ -124,9 +124,11 @@ std::string DescribeTensor(const PaddleTensor &tensor) { void PrintTime(int batch_size, int repeat, int num_threads, int tid, double latency) { + LOG(INFO) << "====================================="; LOG(INFO) << "batch_size: " << batch_size << ", repeat: " << repeat << ", threads: " << num_threads << ", thread id: " << tid << ", latency: " << latency << "ms"; + LOG(INFO) << "====================================="; } } // namespace inference -- GitLab