diff --git a/paddle/fluid/inference/tests/api/CMakeLists.txt b/paddle/fluid/inference/tests/api/CMakeLists.txt index b0f7dcc0dffcb71a8b88c764dbced05fde36745e..aa3da397ff67dd06dd750d336a49056baedaaab6 100644 --- a/paddle/fluid/inference/tests/api/CMakeLists.txt +++ b/paddle/fluid/inference/tests/api/CMakeLists.txt @@ -128,10 +128,10 @@ inference_analysis_api_test_with_fake_data(test_analyzer_resnet50 inference_analysis_api_test_with_fake_data(test_analyzer_mobilenet_depthwise_conv "${INFERENCE_DEMO_INSTALL_DIR}/mobilenet_depthwise_conv" analyzer_resnet50_tester.cc "mobilenet_model.tar.gz" SERIAL) -# bert -set(BERT_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/bert") -download_model_and_data(${BERT_INSTALL_DIR} "bert_model.tar.gz" "bert_data.txt.tar.gz") -inference_analysis_api_test(test_analyzer_bert ${BERT_INSTALL_DIR} analyzer_bert_tester.cc) +# bert, max_len=20 +set(BERT_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/bert20") +download_model_and_data(${BERT_INSTALL_DIR} "bert_model.tar.gz" "bert_data_len20.txt.tar.gz") +inference_analysis_api_test(test_analyzer_bert ${BERT_INSTALL_DIR} analyzer_bert_tester.cc SERIAL) # anakin if (WITH_ANAKIN AND WITH_MKL) # only needed in CI diff --git a/paddle/fluid/inference/tests/api/analyzer_bert_tester.cc b/paddle/fluid/inference/tests/api/analyzer_bert_tester.cc index 24cbd39ea0fbf425a8fed66f6413008f1a97408d..f646fd6d91c81b6738e4fc5278739307fa5f99b5 100644 --- a/paddle/fluid/inference/tests/api/analyzer_bert_tester.cc +++ b/paddle/fluid/inference/tests/api/analyzer_bert_tester.cc @@ -18,7 +18,6 @@ namespace paddle { namespace inference { using paddle::PaddleTensor; -using paddle::contrib::AnalysisConfig; template void GetValueFromStream(std::stringstream *ss, T *t) { @@ -158,12 +157,10 @@ bool LoadInputData(std::vector> *inputs) { return true; } -void SetConfig(contrib::AnalysisConfig *config) { - config->SetModel(FLAGS_infer_model); -} +void SetConfig(AnalysisConfig *config) { config->SetModel(FLAGS_infer_model); } void profile(bool use_mkldnn = false) { - contrib::AnalysisConfig config; + AnalysisConfig config; SetConfig(&config); if (use_mkldnn) { @@ -213,17 +210,14 @@ TEST(Analyzer_bert, compare_mkldnn) { compare(true /* use_mkldnn */); } #endif // Compare Deterministic result -// TODO(luotao): Since each unit-test on CI only have 10 minutes, cancel this to -// decrease the CI time. -// TEST(Analyzer_bert, compare_determine) { -// AnalysisConfig cfg; -// SetConfig(&cfg); -// -// std::vector> inputs; -// LoadInputData(&inputs); -// CompareDeterministic(reinterpret_cast(&cfg), -// inputs); -// } +TEST(Analyzer_bert, compare_determine) { + AnalysisConfig cfg; + SetConfig(&cfg); + + std::vector> inputs; + LoadInputData(&inputs); + CompareDeterministic(reinterpret_cast(&cfg), + inputs); +} } // namespace inference } // namespace paddle diff --git a/paddle/fluid/inference/tests/api/analyzer_rnn1_tester.cc b/paddle/fluid/inference/tests/api/analyzer_rnn1_tester.cc index 5ab8577050df485d0c6bbeffc1d2031c2a00db85..c27c39f40a2067dd2bd2150e4b1e53eab7cdf06e 100644 --- a/paddle/fluid/inference/tests/api/analyzer_rnn1_tester.cc +++ b/paddle/fluid/inference/tests/api/analyzer_rnn1_tester.cc @@ -20,7 +20,6 @@ namespace paddle { namespace inference { using namespace framework; // NOLINT -using namespace contrib; // NOLINT struct DataRecord { std::vector>> link_step_data_all;