diff --git a/paddle/fluid/inference/tests/api/CMakeLists.txt b/paddle/fluid/inference/tests/api/CMakeLists.txt index 88e632bf9db9edf400f3e8397cd30e7c29b9abea..2ca84c80058b35840aff5d072cdc99ecf5165f8e 100644 --- a/paddle/fluid/inference/tests/api/CMakeLists.txt +++ b/paddle/fluid/inference/tests/api/CMakeLists.txt @@ -82,26 +82,6 @@ inference_analysis_api_test(test_analyzer_ocr ${OCR_INSTALL_DIR} analyzer_vis_te inference_analysis_api_test_with_fake_data(test_analyzer_resnet50 "${INFERENCE_DEMO_INSTALL_DIR}/resnet50" analyzer_resnet50_tester.cc "resnet50_model.tar.gz") -# face -set(FACE_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/face") -inference_analysis_api_test_with_fake_data(test_analyzer_face_align1 - "${FACE_INSTALL_DIR}/align1" analyzer_face_tester.cc "face%2Falign1_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_align2 - "${FACE_INSTALL_DIR}/align2" analyzer_face_tester.cc "face%2Falign2_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_feature1 - "${FACE_INSTALL_DIR}/feature1" analyzer_face_tester.cc "face%2Ffeature_id_model.tar.gz") -# TODO(luotao): Disable this test due to analysis is timeout 10 minutes. -# inference_analysis_api_test_with_fake_data(test_analyzer_face_feature2 -# "${FACE_INSTALL_DIR}/feature2" analyzer_face_tester.cc "face%2Ffeature_life_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_detect - "${FACE_INSTALL_DIR}/detect" analyzer_face_tester.cc "face%2Fdetect_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_demark - "${FACE_INSTALL_DIR}/demark" analyzer_face_tester.cc "face%2Fdemark_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_score - "${FACE_INSTALL_DIR}/score" analyzer_face_tester.cc "face%2Fscore_model.tar.gz") -inference_analysis_api_test_with_fake_data(test_analyzer_face_super_res - "${FACE_INSTALL_DIR}/super_res" analyzer_face_tester.cc "face%2Fsuper_res_model.tar.gz") - # anakin if (WITH_ANAKIN AND WITH_MKL) # only needed in CI # anakin rnn1 diff --git a/paddle/fluid/inference/tests/api/analyzer_face_tester.cc b/paddle/fluid/inference/tests/api/analyzer_face_tester.cc deleted file mode 100644 index b7db8887d59d5ce707701b6a4081266b8da881f2..0000000000000000000000000000000000000000 --- a/paddle/fluid/inference/tests/api/analyzer_face_tester.cc +++ /dev/null @@ -1,69 +0,0 @@ -/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - -http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. */ - -#include -#include -#include "paddle/fluid/inference/tests/api/tester_helper.h" - -namespace paddle { -namespace inference { -namespace analysis { - -void SetConfig(AnalysisConfig *cfg) { - cfg->param_file = FLAGS_infer_model + "/params"; - cfg->prog_file = FLAGS_infer_model + "/model"; - cfg->use_gpu = false; - cfg->device = 0; - cfg->enable_ir_optim = true; - cfg->specify_input_name = true; -} - -void SetInput(std::vector> *inputs) { - SetFakeImageInput(inputs, FLAGS_infer_model); -} - -// Easy for profiling independently. -TEST(Analyzer_face, profile) { - AnalysisConfig cfg; - SetConfig(&cfg); - std::vector outputs; - - std::vector> input_slots_all; - SetInput(&input_slots_all); - TestPrediction(cfg, input_slots_all, &outputs, FLAGS_num_threads); -} - -// Check the fuse status -TEST(Analyzer_face, fuse_statis) { - AnalysisConfig cfg; - SetConfig(&cfg); - int num_ops; - auto predictor = CreatePaddlePredictor(cfg); - auto fuse_statis = GetFuseStatis( - static_cast(predictor.get()), &num_ops); -} - -// Compare result of NativeConfig and AnalysisConfig -TEST(Analyzer_face, compare) { - AnalysisConfig cfg; - SetConfig(&cfg); - - std::vector> input_slots_all; - SetInput(&input_slots_all); - CompareNativeAndAnalysis(cfg, input_slots_all); -} - -} // namespace analysis -} // namespace inference -} // namespace paddle diff --git a/paddle/fluid/inference/tests/api/analyzer_resnet50_tester.cc b/paddle/fluid/inference/tests/api/analyzer_resnet50_tester.cc index cd04d888a503aaa53efbcd76ae98d5e5242e1a3f..e5c8dfd22a006d5271248c5b083aab2c22417502 100644 --- a/paddle/fluid/inference/tests/api/analyzer_resnet50_tester.cc +++ b/paddle/fluid/inference/tests/api/analyzer_resnet50_tester.cc @@ -43,13 +43,6 @@ void profile(bool use_mkldnn = false) { std::vector> input_slots_all; SetInput(&input_slots_all); TestPrediction(cfg, input_slots_all, &outputs, FLAGS_num_threads); - - if (FLAGS_num_threads == 1 && !FLAGS_test_all_data) { - PADDLE_ENFORCE_EQ(outputs.size(), 1UL); - size_t size = GetSize(outputs[0]); - // output is a 512-dimension feature - EXPECT_EQ(size, 512 * FLAGS_batch_size); - } } TEST(Analyzer_resnet50, profile) { profile(); } @@ -65,8 +58,7 @@ TEST(Analyzer_resnet50, fuse_statis) { auto predictor = CreatePaddlePredictor(cfg); auto fuse_statis = GetFuseStatis( static_cast(predictor.get()), &num_ops); - ASSERT_TRUE(fuse_statis.count("fc_fuse")); - EXPECT_EQ(fuse_statis.at("fc_fuse"), 1); + LOG(INFO) << "num_ops: " << num_ops; } // Compare result of NativeConfig and AnalysisConfig