# In Windows, c_api test link must link both 2 shared to avoid symbols redefinition,
# in Linux, c_api test cant do like this or graph_to_program register more than once.
# Both Windows and Linux can only use paddle_inference_c, but this will increase size
# of build folder by 30G.
set(inference_api_tester_deps paddle_inference_api analysis_config)

cc_test(
  test_paddle_inference_api
  SRCS api_tester.cc
  DEPS ${inference_api_tester_deps})

cc_test(
  inference_api_helper_test
  SRCS helper_test.cc
  DEPS ${inference_api_tester_deps})

if(WITH_ONNXRUNTIME AND WIN32)
  # Copy onnxruntime for some c++ test in Windows, since the test will
  # be build only in CI, so suppose the generator in Windows is Ninja.
  copy_onnx(test_paddle_inference_api)
endif()

if(WITH_TESTING AND WITH_INFERENCE_API_TEST)
  if(WIN32)
    set(INFERENCE_C_EXTRA_DEPS paddle_inference_shared
                               paddle_inference_c_shared)
  else()
    set(INFERENCE_C_EXTRA_DEPS paddle_inference_shared paddle_inference_c)
  endif()

  function(download_data install_dir data_file check_sum)
    string(REGEX MATCH "[^/\\]+$" file_name ${data_file})
    if(NOT EXISTS ${install_dir}/${file_name})
      inference_download_and_uncompress(${install_dir} ${INFERENCE_URL}
                                        ${data_file} ${check_sum})
    endif()
  endfunction()

  function(download_data_without_verify install_dir data_file)
    string(REGEX MATCH "[^/\\]+$" file_name ${data_file})
    if(NOT EXISTS ${install_dir}/${file_name})
      inference_download_and_uncompress_without_verify(
        ${install_dir} ${INFERENCE_URL} ${data_file})
    endif()
  endfunction()

  function(download_int8_data install_dir data_file check_sum)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress(${install_dir} ${INFERENCE_URL}/int8
                                        ${data_file} ${check_sum})
    endif()
  endfunction()

  function(download_int8_data_without_verify install_dir data_file)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress_without_verify(
        ${install_dir} ${INFERENCE_URL}/int8 ${data_file})
    endif()
  endfunction()

  function(download_bfloat16_data install_dir data_file check_sum)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress(
        ${install_dir} ${INFERENCE_URL}/bfloat16 ${data_file} ${check_sum})
    endif()
  endfunction()

  function(download_bfloat16_data_without_verify install_dir data_file)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress_without_verify(
        ${install_dir} ${INFERENCE_URL}/bfloat16 ${data_file})
    endif()
  endfunction()

  function(download_GRU_data install_dir data_file check_sum)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress(${install_dir} ${INFERENCE_URL}/gru
                                        ${data_file} ${check_sum})
    endif()
  endfunction()

  function(download_GRU_data_without_verify install_dir data_file)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress_without_verify(
        ${install_dir} ${INFERENCE_URL}/gru ${data_file})
    endif()
  endfunction()

  function(download_quant_data install_dir data_file check_sum)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress(
        ${install_dir} ${INFERENCE_URL}/int8/QAT_models ${data_file}
        ${check_sum})
    endif()
  endfunction()

  function(download_quant_data_without_verify install_dir data_file)
    if(NOT EXISTS ${install_dir}/${data_file})
      inference_download_and_uncompress_without_verify(
        ${install_dir} ${INFERENCE_URL}/int8/QAT_models ${data_file})
    endif()
  endfunction()

  function(download_model_and_data install_dir model_name model_check_sum
           data_name data_check_sum)
    download_data(${install_dir} ${model_name} ${model_check_sum})
    download_data(${install_dir} ${data_name} ${data_check_sum})
  endfunction()

  function(download_model_and_data_without_verify install_dir model_name
           data_name)
    download_data_without_verify(${install_dir} ${model_name})
    download_data_without_verify(${install_dir} ${data_name})
  endfunction()

  function(download_result install_dir result_name check_sum)
    download_data(${install_dir} ${result_name} ${check_sum})
  endfunction()

  function(download_result_without_verify install_dir result_name)
    download_data_without_verify(${install_dir} ${result_name})
  endfunction()

  function(inference_analysis_api_test target install_dir filename)
    inference_analysis_test(
      ${target}
      SRCS
      ${filename}
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${install_dir}/model
      --infer_data=${install_dir}/data.txt
      --refer_result=${install_dir}/result.txt)
  endfunction()

  function(inference_analysis_api_int8_test target install_dir filename)
    inference_analysis_test(
      ${target}
      SRCS
      ${filename}
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${install_dir}/model
      --infer_data=${install_dir}/data.txt
      --refer_result=${install_dir}/result.txt
      --accuracy=0.8
      --batch_size=5
      --enable_int8_ptq=true)
  endfunction()

  function(inference_multiple_models_analysis_api_test target install_dir
           filename)
    inference_analysis_test(
      ${target}
      SRCS
      ${filename}
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${install_dir}/mobilenet_v2_models/1
      --infer_model2=${install_dir}/mobilenet_v2_models/xx
      --infer_model3=${install_dir}/mobilenet_v2_models/3)
  endfunction()

  function(inference_analysis_api_test_build TARGET_NAME filename)
    inference_analysis_test_build(${TARGET_NAME} SRCS ${filename} EXTRA_DEPS
                                  paddle_inference_shared python)
  endfunction()

  function(inference_analysis_api_int8_test_run TARGET_NAME test_binary
           model_dir data_path)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${model_dir}/model
      --infer_data=${data_path}
      --warmup_batch_size=${WARMUP_BATCH_SIZE}
      --batch_size=50
      --enable_int8_ptq=true
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --iterations=2)
  endfunction()

  function(inference_analysis_api_int8_test_run_custom_warmup_batch_size
           TARGET_NAME test_binary model_dir data_path warmup_batch_size)
    set(WARMUP_BATCH_SIZE ${warmup_batch_size})
    inference_analysis_api_int8_test_run(${TARGET_NAME} ${test_binary}
                                         ${model_dir} ${data_path})
  endfunction()

  function(inference_analysis_api_bfloat16_test_run TARGET_NAME test_binary
           model_dir data_path)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${model_dir}/model
      --infer_data=${data_path}
      --batch_size=50
      --enable_bf16=true
      --paddle_num_threads=${CPU_NUM_THREADS_ON_CI}
      --iterations=2)
  endfunction()

  function(inference_analysis_api_object_dection_int8_test_run TARGET_NAME
           test_binary model_dir data_path)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${model_dir}/model
      --infer_data=${data_path}
      --warmup_batch_size=10
      --batch_size=300
      --enable_int8_ptq=true
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --iterations=1)
  endfunction()

  function(inference_analysis_api_test_with_fake_data_build TARGET_NAME
           filename)
    inference_analysis_test_build(${TARGET_NAME} SRCS ${filename} EXTRA_DEPS
                                  paddle_inference_shared python)
  endfunction()

  function(inference_analysis_api_test_with_fake_data_run TARGET_NAME
           test_binary model_dir disable_fc)
    inference_analysis_test_run(
      ${TARGET_NAME} COMMAND ${test_binary} ARGS
      --infer_model=${model_dir}/model --disable_mkldnn_fc=${disable_fc})
  endfunction()

  function(
    inference_analysis_api_quant_test_run
    TARGET_NAME
    test_binary
    fp32_model_dir
    int8_model_dir
    data_path
    enable_int8_qat)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --fp32_model=${fp32_model_dir}
      --int8_model=${int8_model_dir}
      --infer_data=${data_path}
      --batch_size=50
      --enable_int8_qat=${enable_int8_qat}
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --with_accuracy_layer=false
      --iterations=2)
  endfunction()

  function(inference_analysis_api_lexical_test_run TARGET_NAME test_binary
           infer_model data_path)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${infer_model}
      --infer_data=${data_path}
      --batch_size=50
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --with_accuracy_layer=true
      --use_analysis=true
      --iterations=2)
  endfunction()

  function(inference_analysis_api_lexical_bfloat16_test_run TARGET_NAME
           test_binary infer_model data_path)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${infer_model}
      --infer_data=${data_path}
      --batch_size=50
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --with_accuracy_layer=true
      --use_analysis=true
      --enable_bf16=true
      --iterations=2)
  endfunction()

  function(
    inference_analysis_api_lexical_int8_test_run
    TARGET_NAME
    test_binary
    infer_model
    data_path
    enable_int8_ptq
    enable_int8_qat
    fuse_multi_gru)
    inference_analysis_test_run(
      ${TARGET_NAME}
      COMMAND
      ${test_binary}
      ARGS
      --infer_model=${infer_model}
      --infer_data=${data_path}
      --batch_size=100
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --with_accuracy_layer=true
      --use_analysis=true
      --enable_int8_ptq=${enable_int8_ptq}
      --enable_int8_qat=${enable_int8_qat}
      --quantized_accuracy=0.015
      --fuse_multi_gru=${fuse_multi_gru}
      --iterations=4)
  endfunction()

  function(preprocess_data2bin_test_run target py_script_source data_dir
           output_file)
    py_test(${target}
            SRCS ${CMAKE_CURRENT_SOURCE_DIR}/${py_script_source} ARGS
                 --data_dir=${data_dir} --output_file=${output_file} --local)
  endfunction()

  if(NOT APPLE AND WITH_MKLML)
    # RNN1
    set(RNN1_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/rnn1")
    download_model_and_data_without_verify(
      ${RNN1_INSTALL_DIR} "rnn1/model.tar.gz" "rnn1/data.txt.tar.gz")
    inference_analysis_api_test(test_analyzer_rnn1 ${RNN1_INSTALL_DIR}
                                analyzer_rnn1_tester.cc)

    # seq_pool1
    set(SEQ_POOL1_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/seq_pool")
    download_model_and_data_without_verify(
      ${SEQ_POOL1_INSTALL_DIR} "seq_pool1_model_.tar.gz"
      "seq_pool1_data.txt.tar.gz")
    inference_analysis_api_test(
      test_analyzer_seq_pool1_compare_determine ${SEQ_POOL1_INSTALL_DIR}
      analyzer_seq_pool1_compare_determine_tester.cc)
    inference_analysis_api_test(
      test_analyzer_seq_pool1 ${SEQ_POOL1_INSTALL_DIR}
      analyzer_seq_pool1_compare_tester.cc)
    inference_analysis_api_test(
      test_analyzer_seq_pool1_fuse_compare_zero_copy ${SEQ_POOL1_INSTALL_DIR}
      analyzer_seq_pool1_fuse_compare_zero_copy_tester.cc)
    inference_analysis_api_test(
      test_analyzer_seq_pool1_fuse_statis ${SEQ_POOL1_INSTALL_DIR}
      analyzer_seq_pool1_fuse_statis_tester.cc)
    inference_analysis_api_test(
      test_analyzer_seq_pool1_profile ${SEQ_POOL1_INSTALL_DIR}
      analyzer_seq_pool1_profile_tester.cc)
    if(NOT WIN32)
      set_tests_properties(test_analyzer_seq_pool1_compare_determine
                           PROPERTIES TIMEOUT 120)
      set_tests_properties(test_analyzer_seq_pool1 PROPERTIES TIMEOUT 120)
      set_tests_properties(test_analyzer_seq_pool1_fuse_compare_zero_copy
                           PROPERTIES TIMEOUT 120)
      set_tests_properties(test_analyzer_seq_pool1_fuse_statis
                           PROPERTIES TIMEOUT 120)
      set_tests_properties(test_analyzer_seq_pool1_profile PROPERTIES TIMEOUT
                                                                      120)
    endif()
  else()
    # TODO: fix this test on MACOS and OPENBLAS, the reason is that
    # fusion_seqexpand_concat_fc_op is not supported on MACOS and OPENBLAS
    message(
      WARNING
        "These tests has been disabled in OSX or WITH_MKL=OFF before being fixed: \n test_analyzer_rnn1"
    )
    message(
      WARNING
        "These tests has been disabled in OSX or WITH_MKL=OFF before being fixed: \n test_analyzer_seq_pool1"
    )
  endif()

  # RNN2
  set(RNN2_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/rnn2")
  download_model_and_data_without_verify(
    ${RNN2_INSTALL_DIR} "rnn2_model.tar.gz" "rnn2_data.txt.tar.gz")
  inference_analysis_api_test(test_analyzer_rnn2 ${RNN2_INSTALL_DIR}
                              analyzer_rnn2_tester.cc)

  # TODO(luotao, Superjom) Disable DAM test, temporarily fix
  # https://github.com/PaddlePaddle/Paddle/issues/15032#issuecomment-455990914.
  # After inference framework refactor, will reopen it.
  # normal DAM
  set(DAM_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/dam")
  download_model_and_data_without_verify(${DAM_INSTALL_DIR} "DAM_model.tar.gz"
                                         "DAM_data.txt.tar.gz")
  #inference_analysis_api_test(test_analyzer_dam ${DAM_INSTALL_DIR} analyzer_dam_tester.cc EXTRA_DEPS legacy_allocator)

  # small DAM
  set(DAM_SMALL_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/small_dam")
  download_model_and_data_without_verify(
    ${DAM_SMALL_INSTALL_DIR} "dam_small_model.tar.gz"
    "dam_small_data.txt.tar.gz")
  inference_analysis_test(
    test_analyzer_small_dam
    SRCS
    analyzer_dam_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${DAM_SMALL_INSTALL_DIR}/model
    --infer_data=${DAM_SMALL_INSTALL_DIR}/data.txt)

  #save model
  inference_analysis_api_test(test_analyzer_save_model ${DAM_SMALL_INSTALL_DIR}
                              analyzer_save_model_tester.cc)

  # chinese_ner
  set(CHINESE_NER_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/chinese_ner")
  download_model_and_data_without_verify(
    ${CHINESE_NER_INSTALL_DIR} "chinese_ner_model.tar.gz"
    "chinese_ner-data.txt.tar.gz")
  inference_analysis_api_test(test_analyzer_ner ${CHINESE_NER_INSTALL_DIR}
                              analyzer_ner_tester.cc)

  # lac
  set(LAC_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/lac")
  download_model_and_data(
    ${LAC_INSTALL_DIR} "lac_model.tar.gz" 419ca6eb85f57a01bfe173591910aec5
    "lac_data.txt.tar.gz" 9983539cd6b34fbdc411e43422776bfd)
  inference_analysis_api_test(test_analyzer_lac ${LAC_INSTALL_DIR}
                              analyzer_lac_tester.cc)

  # Ernie
  set(ERNIE_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/Ernie")
  download_model_and_data(
    ${ERNIE_INSTALL_DIR} "Ernie_model.tar.gz" aa59192dd41ed377f9f168e3a1309fa6
    "Ernie_data.txt.tar.gz" 5396e63548edad7ca561e7e26a9476d1)
  download_result(${ERNIE_INSTALL_DIR} "Ernie_result.txt.tar.gz"
                  73beea65abda2edb61c1662cd3180c62)
  if(WITH_GPU)
    inference_analysis_api_test(test_analyzer_ernie ${ERNIE_INSTALL_DIR}
                                analyzer_ernie_tester.cc)
    inference_analysis_api_test(gpu_ernie_half_test ${ERNIE_INSTALL_DIR}
                                gpu_ernie_half_test.cc)
    set_tests_properties(gpu_ernie_half_test PROPERTIES TIMEOUT 60)
  endif()
  inference_analysis_api_int8_test(
    test_analyzer_ernie_int8 ${ERNIE_INSTALL_DIR} analyzer_ernie_int8_tester.cc)

  # Ernie large
  set(ERNIE_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/Ernie_Large")
  download_model_and_data(
    ${ERNIE_INSTALL_DIR} "Ernie_large_model.tar.gz"
    af7715245ed32cc77374625d4c80f7ef "Ernie_large_data.txt.tar.gz"
    edb2113eec93783cad56ed76d47ba57f)
  download_result(${ERNIE_INSTALL_DIR} "Ernie_large_result.txt.tar.gz"
                  1facda98eef1085dc9d435ebf3f23a73)
  inference_analysis_test(
    test_analyzer_ernie_large
    SRCS
    analyzer_ernie_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${ERNIE_INSTALL_DIR}/model
    --infer_data=${ERNIE_INSTALL_DIR}/data.txt
    --refer_result=${ERNIE_INSTALL_DIR}/result.txt
    --ernie_large=true)
  if(NOT WIN32
     AND NOT APPLE
     AND TEST test_analyzer_ernie_large)
    set_tests_properties(test_analyzer_ernie_large
                         PROPERTIES TIMEOUT 150 LABELS "RUN_TYPE=NIGHTLY")
  endif()
  if(WIN32 AND TEST test_analyzer_ernie_large)
    set_tests_properties(test_analyzer_ernie_large PROPERTIES TIMEOUT 200)
  endif()

  # text_classification
  set(TEXT_CLASSIFICATION_INSTALL_DIR
      "${INFERENCE_DEMO_INSTALL_DIR}/text_classification")
  download_model_and_data(
    ${TEXT_CLASSIFICATION_INSTALL_DIR} "text-classification-Senta.tar.gz"
    3f0f440313ca50e26184e65ffd5809ab "text_classification_data.txt.tar.gz"
    36ae620020cc3377f45ed330dd36238f)
  inference_analysis_api_test(
    test_analyzer_text_classification ${TEXT_CLASSIFICATION_INSTALL_DIR}
    analyzer_text_classification_tester.cc)

  # seq_conv1
  set(SEQ_CONV1_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/seq_conv1")
  download_model_and_data_without_verify(
    ${SEQ_CONV1_INSTALL_DIR} "seq_conv1_model.tar.gz"
    "seq_conv1_data.txt.tar.gz")
  inference_analysis_api_test(test_analyzer_seq_conv1 ${SEQ_CONV1_INSTALL_DIR}
                              analyzer_seq_conv1_tester.cc)

  # transformer, the dataset only works on batch_size=8 now
  set(TRANSFORMER_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/transformer")
  download_model_and_data_without_verify(
    ${TRANSFORMER_INSTALL_DIR} "temp/transformer_model.tar.gz"
    "temp/transformer_data.txt.tar.gz")
  inference_analysis_test(
    test_analyzer_transformer
    SRCS
    analyzer_transformer_compare_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${TRANSFORMER_INSTALL_DIR}/model
    --infer_data=${TRANSFORMER_INSTALL_DIR}/data.txt
    --batch_size=8
    --cpu_num_threads=${CPU_NUM_THREADS_ON_CI})
  inference_analysis_test(
    test_analyzer_transformer_fuse
    SRCS
    analyzer_transformer_fuse_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${TRANSFORMER_INSTALL_DIR}/model
    --infer_data=${TRANSFORMER_INSTALL_DIR}/data.txt
    --batch_size=8
    --cpu_num_threads=${CPU_NUM_THREADS_ON_CI})
  inference_analysis_test(
    test_analyzer_transformer_profile
    SRCS
    analyzer_transformer_profile_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${TRANSFORMER_INSTALL_DIR}/model
    --infer_data=${TRANSFORMER_INSTALL_DIR}/data.txt
    --batch_size=8
    --cpu_num_threads=${CPU_NUM_THREADS_ON_CI})

  # VIT-OCR
  set(VIT_OCR_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/vit")
  if(NOT EXISTS ${VIT_OCR_INSTALL_DIR}/vit_ocr.tgz)
    inference_download_and_uncompress_without_verify(
      ${VIT_OCR_INSTALL_DIR} ${INFERENCE_URL} "ocr/vit_ocr.tgz")
  endif()
  inference_analysis_test(
    test_analyzer_vit_ocr
    SRCS
    analyzer_vit_ocr_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${VIT_OCR_INSTALL_DIR}/vit_ocr/model
    --infer_data=${VIT_OCR_INSTALL_DIR}/vit_ocr/datavit.txt)

  # ocr
  set(OCR_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/ocr")
  if(NOT EXISTS ${OCR_INSTALL_DIR}/ocr.tar.gz)
    inference_download_and_uncompress_without_verify(
      ${OCR_INSTALL_DIR} "http://paddlemodels.bj.bcebos.com/"
      "inference-vis-demos/ocr.tar.gz")
  endif()
  inference_analysis_api_test(test_analyzer_ocr ${OCR_INSTALL_DIR}
                              analyzer_vis_tester.cc)

  # densebox
  set(DENSEBOX_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/densebox")
  download_data_without_verify(${DENSEBOX_INSTALL_DIR} "densebox.tar.gz")
  inference_analysis_test(
    test_analyzer_detect_functional_mkldnn
    SRCS
    analyzer_detect_functional_mkldnn_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${DENSEBOX_INSTALL_DIR}/model
    --infer_data=${DENSEBOX_INSTALL_DIR}/detect_input_50.txt
    --infer_shape=${DENSEBOX_INSTALL_DIR}/shape_50.txt)

  # mobilenet with transpose op
  set(MOBILENET_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/mobilenet")
  if(NOT EXISTS ${MOBILENET_INSTALL_DIR}/mobilenet.tar.gz)
    inference_download_and_uncompress_without_verify(
      ${MOBILENET_INSTALL_DIR} "http://paddlemodels.bj.bcebos.com/"
      "inference-vis-demos/mobilenet.tar.gz")
  endif()
  inference_analysis_api_test(test_analyzer_mobilenet_transpose
                              ${MOBILENET_INSTALL_DIR} analyzer_vis_tester.cc)

  ### Image classification tests with fake data
  set(IMG_CLASS_TEST_APP "test_analyzer_image_classification")
  set(IMG_CLASS_TEST_APP_SRC "analyzer_image_classification_tester.cc")

  # build test binary to be used in subsequent tests
  inference_analysis_api_test_with_fake_data_build(${IMG_CLASS_TEST_APP}
                                                   ${IMG_CLASS_TEST_APP_SRC})

  # googlenet
  set(GOOGLENET_MODEL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/googlenet")
  download_data_without_verify(${GOOGLENET_MODEL_DIR} "googlenet.tar.gz")
  inference_analysis_api_test_with_fake_data_run(
    test_analyzer_googlenet ${IMG_CLASS_TEST_APP} ${GOOGLENET_MODEL_DIR} false)

  # resnet50
  set(RESNET50_MODEL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/resnet50")
  download_data_without_verify(${RESNET50_MODEL_DIR} "resnet50_model.tar.gz")
  inference_analysis_api_test_with_fake_data_run(
    test_analyzer_resnet50 ${IMG_CLASS_TEST_APP} ${RESNET50_MODEL_DIR} true)
  if(WIN32)
    set_tests_properties(test_analyzer_resnet50 PROPERTIES TIMEOUT 200)
  endif()

  # mobilenet with depthwise_conv op
  set(MOBILENET_MODEL_DIR
      "${INFERENCE_DEMO_INSTALL_DIR}/mobilenet_depthwise_conv")
  download_data_without_verify(${MOBILENET_MODEL_DIR} "mobilenet_model.tar.gz")
  inference_analysis_api_test_with_fake_data_run(
    test_analyzer_mobilenet_depthwise_conv ${IMG_CLASS_TEST_APP}
    ${MOBILENET_MODEL_DIR} false)

  if(WITH_MKLDNN)

    ### INT8 tests

    set(INT8_DATA_DIR "${INFERENCE_DEMO_INSTALL_DIR}/int8v2")

    ## Image classification models

    # ImageNet small dataset
    # It may be already downloaded for Quant & INT8 unit tests
    set(IMAGENET_DATA_ARCHIVE "imagenet_val_100_tail.tar.gz")
    set(IMAGENET_DATA_DIR "${INFERENCE_DEMO_INSTALL_DIR}/imagenet")
    set(IMAGENET_DATA_PATH "${IMAGENET_DATA_DIR}/data.bin")
    download_int8_data_without_verify(${IMAGENET_DATA_DIR}
                                      ${IMAGENET_DATA_ARCHIVE})

    # build test binary to be used in subsequent tests
    set(INT8_IMG_CLASS_TEST_APP "test_analyzer_int8_image_classification")
    set(INT8_IMG_CLASS_TEST_APP_SRC
        "analyzer_int8_image_classification_tester.cc")
    inference_analysis_api_test_build(${INT8_IMG_CLASS_TEST_APP}
                                      ${INT8_IMG_CLASS_TEST_APP_SRC})

    # resnet50 int8
    set(INT8_RESNET50_MODEL_DIR "${INT8_DATA_DIR}/resnet50")
    download_int8_data_without_verify(${INT8_RESNET50_MODEL_DIR}
                                      "resnet50_int8_model.tar.gz")
    inference_analysis_api_int8_test_run(
      test_analyzer_int8_resnet50 ${INT8_IMG_CLASS_TEST_APP}
      ${INT8_RESNET50_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # mobilenetv1 int8
    set(INT8_MOBILENETV1_MODEL_DIR "${INT8_DATA_DIR}/mobilenetv1")
    download_int8_data_without_verify(${INT8_MOBILENETV1_MODEL_DIR}
                                      "mobilenetv1_int8_model.tar.gz")
    inference_analysis_api_int8_test_run(
      test_analyzer_int8_mobilenetv1 ${INT8_IMG_CLASS_TEST_APP}
      ${INT8_MOBILENETV1_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # mobilenetv2 int8
    set(INT8_MOBILENETV2_MODEL_DIR "${INT8_DATA_DIR}/mobilenetv2")
    download_int8_data_without_verify(${INT8_MOBILENETV2_MODEL_DIR}
                                      "mobilenet_v2_int8_model.tar.gz")
    inference_analysis_api_int8_test_run(
      test_analyzer_int8_mobilenetv2 ${INT8_IMG_CLASS_TEST_APP}
      ${INT8_MOBILENETV2_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # resnet101 int8
    set(INT8_RESNET101_MODEL_DIR "${INT8_DATA_DIR}/resnet101")
    download_int8_data_without_verify(${INT8_RESNET101_MODEL_DIR}
                                      "Res101_int8_model.tar.gz")
    #   inference_analysis_api_int8_test_run(test_analyzer_int8_resnet101 ${INT8_IMG_CLASS_TEST_APP} ${INT8_RESNET101_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # vgg16 int8
    set(INT8_VGG16_MODEL_DIR "${INT8_DATA_DIR}/vgg16")
    download_int8_data_without_verify(${INT8_VGG16_MODEL_DIR}
                                      "VGG16_int8_model.tar.gz")
    #  inference_analysis_api_int8_test_run(test_analyzer_int8_vgg16 ${INT8_IMG_CLASS_TEST_APP} ${INT8_VGG16_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # vgg19 int8
    set(INT8_VGG19_MODEL_DIR "${INT8_DATA_DIR}/vgg19")
    download_int8_data_without_verify(${INT8_VGG19_MODEL_DIR}
                                      "VGG19_int8_model.tar.gz")
    #   inference_analysis_api_int8_test_run(test_analyzer_int8_vgg19 ${INT8_IMG_CLASS_TEST_APP} ${INT8_VGG19_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # googlenet int8
    set(INT8_GOOGLENET_MODEL_DIR "${INT8_DATA_DIR}/googlenet")
    download_int8_data_without_verify(${INT8_GOOGLENET_MODEL_DIR}
                                      "GoogleNet_int8_model.tar.gz")
    inference_analysis_api_int8_test_run_custom_warmup_batch_size(
      test_analyzer_int8_googlenet ${INT8_IMG_CLASS_TEST_APP}
      ${INT8_GOOGLENET_MODEL_DIR} ${IMAGENET_DATA_PATH} 10)

    # mobilenetv3_large_x1_0 int8
    set(INT8_MOBILENETV3_LARGE_MODEL_DIR "${INT8_DATA_DIR}/mobilenetv3_large")
    set(INT8_MOBILENETV3_FILE_NAME "MobileNetV3_large_x1_0_infer.tar")
    if(NOT EXISTS
       ${INT8_MOBILENETV3_LARGE_MODEL_DIR}/${INT8_MOBILENETV3_FILE_NAME})
      inference_download_and_uncompress_without_verify(
        ${INT8_MOBILENETV3_LARGE_MODEL_DIR}
        "https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/inference/"
        ${INT8_MOBILENETV3_FILE_NAME})
    endif()
    inference_analysis_test_run(
      test_analyzer_int8_mobilenetv3_large
      COMMAND
      ${INT8_IMG_CLASS_TEST_APP}
      ARGS
      --infer_model=${INT8_MOBILENETV3_LARGE_MODEL_DIR}/MobileNetV3_large_x1_0_infer
      --infer_data=${IMAGENET_DATA_PATH}
      --warmup_batch_size=50
      --batch_size=1
      --enable_int8_ptq=true
      --cpu_num_threads=${CPU_NUM_THREADS_ON_CI}
      --iterations=100
      --with_accuracy_layer=false)

    ### BFLOAT16 tests

    # build test binary to be used in subsequent tests
    set(BF16_IMG_CLASS_TEST_APP "test_analyzer_bfloat16_image_classification")
    set(BF16_IMG_CLASS_TEST_APP_SRC
        "analyzer_bfloat16_image_classification_tester.cc")
    inference_analysis_api_test_build(${BF16_IMG_CLASS_TEST_APP}
                                      ${BF16_IMG_CLASS_TEST_APP_SRC})

    # resnet50 bfloat16
    inference_analysis_api_bfloat16_test_run(
      test_analyzer_bfloat16_resnet50 ${BF16_IMG_CLASS_TEST_APP}
      ${INT8_RESNET50_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # googlenet bfloat16
    inference_analysis_api_bfloat16_test_run(
      test_analyzer_bfloat16_googlenet ${BF16_IMG_CLASS_TEST_APP}
      ${INT8_GOOGLENET_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # mobilenetv1 bfloat16
    inference_analysis_api_bfloat16_test_run(
      test_analyzer_bfloat16_mobilenetv1 ${BF16_IMG_CLASS_TEST_APP}
      ${INT8_MOBILENETV1_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # mobilenetv2 bfloat16
    inference_analysis_api_bfloat16_test_run(
      test_analyzer_bfloat16_mobilenetv2 ${BF16_IMG_CLASS_TEST_APP}
      ${INT8_MOBILENETV2_MODEL_DIR} ${IMAGENET_DATA_PATH})

    # mobilenetv3_large
    inference_analysis_test_run(
      test_analyzer_bfloat16_mobilenetv3_large
      COMMAND
      ${BF16_IMG_CLASS_TEST_APP}
      ARGS
      --infer_model=${INT8_MOBILENETV3_LARGE_MODEL_DIR}/MobileNetV3_large_x1_0_infer
      --infer_data=${IMAGENET_DATA_PATH}
      --batch_size=1
      --enable_bf16=true
      --paddle_num_threads=${CPU_NUM_THREADS_ON_CI}
      --iterations=100
      --with_accuracy_layer=false)

    ### Object detection models
    set(PASCALVOC_DATA_PATH "${INT8_DATA_DIR}/pascalvoc_val_head_300.bin")
    set(INT8_OBJ_DETECT_TEST_APP "test_analyzer_int8_object_detection")
    set(INT8_OBJ_DETECT_TEST_APP_SRC "analyzer_int8_object_detection_tester.cc")

    # download dataset if necessary
    download_int8_data_without_verify(${INT8_DATA_DIR}
                                      "pascalvoc_val_head_300.tar.gz")

    # build test binary to be used in subsequent tests
    inference_analysis_api_test_build(${INT8_OBJ_DETECT_TEST_APP}
                                      ${INT8_OBJ_DETECT_TEST_APP_SRC})

    # mobilenet-ssd int8
    set(INT8_MOBILENET_SSD_MODEL_DIR "${INT8_DATA_DIR}/mobilenet-ssd")
    download_int8_data_without_verify(${INT8_MOBILENET_SSD_MODEL_DIR}
                                      "mobilenet_ssd_int8_model.tar.gz")
    inference_analysis_api_object_dection_int8_test_run(
      test_analyzer_int8_mobilenet_ssd ${INT8_OBJ_DETECT_TEST_APP}
      ${INT8_MOBILENET_SSD_MODEL_DIR} ${PASCALVOC_DATA_PATH})

    ### Lexcial analysis GRU model
    set(GRU_PATH "${INFERENCE_DEMO_INSTALL_DIR}/gru")
    download_gru_data_without_verify("${GRU_PATH}" "GRU_eval_data.tar.gz")
    download_gru_data_without_verify("${GRU_PATH}" "GRU_eval_model_v2.tar.gz")
    set(GRU_DATA_PATH "${GRU_PATH}/GRU_eval_data.bin")
    set(GRU_MODEL_PATH "${GRU_PATH}/GRU_eval_model_v2")
    set(LEXICAL_TEST_APP "test_analyzer_lexical_analysis")
    set(LEXICAL_TEST_APP_SRC "analyzer_lexical_analysis_gru_tester.cc")

    # build test binary to be used in subsequent tests
    inference_analysis_api_test_build(${LEXICAL_TEST_APP}
                                      ${LEXICAL_TEST_APP_SRC})
    # run lexcial analysis test
    inference_analysis_api_lexical_test_run(
      test_analyzer_lexical_gru ${LEXICAL_TEST_APP} ${GRU_MODEL_PATH}
      ${GRU_DATA_PATH})
    # run bfloat16 lexical analysis test
    inference_analysis_api_lexical_bfloat16_test_run(
      test_analyzer_lexical_gru_bfloat16 ${LEXICAL_TEST_APP} ${GRU_MODEL_PATH}
      ${GRU_DATA_PATH})
    # run post-training quantization lexical analysis test
    inference_analysis_api_lexical_int8_test_run(
      test_analyzer_lexical_gru_int8
      ${LEXICAL_TEST_APP}
      ${GRU_MODEL_PATH}
      ${GRU_DATA_PATH}
      true # enable_int8_ptq
      false # enable_int8_qat
      false) # fuse_multi_gru
    # run post-training quantization lexical analysis test with multi_gru fuse
    inference_analysis_api_lexical_int8_test_run(
      test_analyzer_lexical_gru_int8_multi_gru
      ${LEXICAL_TEST_APP}
      ${GRU_MODEL_PATH}
      ${GRU_DATA_PATH}
      true # enable_int8_ptq
      false # enable_int8_qat
      true) # fuse_multi_gru

    # run qat gru test
    set(QAT_GRU_MODEL_ARCHIVE "GRU_quant_acc.tar.gz")
    set(QAT_GRU_MODEL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/quant/GRU_quant2")
    download_quant_data(${QAT_GRU_MODEL_DIR} ${QAT_GRU_MODEL_ARCHIVE}
                        cf207f8076dcfb8b74d8b6bdddf9090c)

    inference_analysis_api_lexical_int8_test_run(
      test_analyzer_lexical_gru_qat_int8
      ${LEXICAL_TEST_APP}
      "${QAT_GRU_MODEL_DIR}/GRU_quant_acc"
      ${GRU_DATA_PATH}
      false # enable_int8_ptq
      true # enable_int8_qat
      false) # fuse_multi_gru

    ### optimized FP32 vs. Quant INT8 tests

    set(QUANT_DATA_DIR "${INFERENCE_DEMO_INSTALL_DIR}/quant")
    set(QUANT_IMG_CLASS_TEST_APP "test_analyzer_quant_image_classification")
    set(QUANT_IMG_CLASS_TEST_APP_SRC
        "analyzer_quant_image_classification_tester.cc")

    # build test binary to be used in subsequent tests
    inference_analysis_api_test_build(${QUANT_IMG_CLASS_TEST_APP}
                                      ${QUANT_IMG_CLASS_TEST_APP_SRC})

    # MobileNetV1 FP32 vs. Quant INT8
    # The FP32 model should already be downloaded for slim Quant unit tests on Linux
    set(QUANT2_MobileNetV1_MODEL_DIR "${QUANT_DATA_DIR}/MobileNetV1_quant2")
    set(QUANT2_INT8_MobileNetV1_MODEL_DIR
        "${QUANT_DATA_DIR}/MobileNetV1_quant2_int8")
    if(NOT LINUX)
      download_quant_data_without_verify(${QUANT2_MobileNetV1_MODEL_DIR}
                                         "MobileNet_qat_perf.tar.gz")
    endif()
    download_quant_data_without_verify(${QUANT2_INT8_MobileNetV1_MODEL_DIR}
                                       "MobileNet_qat_perf_int8.tar.gz")
    inference_analysis_api_quant_test_run(
      test_analyzer_quant_performance_benchmark
      ${QUANT_IMG_CLASS_TEST_APP}
      ${QUANT2_MobileNetV1_MODEL_DIR}/MobileNet_qat_perf/float
      ${QUANT2_INT8_MobileNetV1_MODEL_DIR}/MobileNet_qat_perf_int8
      ${IMAGENET_DATA_PATH}
      false)

    # Quant2 MobileNetV1
    inference_analysis_api_quant_test_run(
      test_analyzer_quant2_mobilenetv1_mkldnn
      ${QUANT_IMG_CLASS_TEST_APP}
      ${QUANT2_MobileNetV1_MODEL_DIR}/MobileNet_qat_perf/float
      ${QUANT2_MobileNetV1_MODEL_DIR}/MobileNet_qat_perf/float
      ${IMAGENET_DATA_PATH}
      true)

    # Quant2 ResNet50 with input/output scales in `fake_quantize_range_abs_max` operators and the `out_threshold` attributes,
    # with weight scales in `fake_channel_wise_dequantize_max_abs` operators
    set(QUANT2_RESNET50_CHANNELWISE_MODEL_DIR
        "${QUANT_DATA_DIR}/ResNet50_quant2_channelwise")
    set(QUANT2_RESNET50_CHANNELWISE_MODEL_ARCHIVE
        "ResNet50_qat_channelwise.tar.gz")
    if(NOT LINUX)
      download_quant_data_without_verify(
        ${QUANT2_RESNET50_CHANNELWISE_MODEL_DIR}
        ${QUANT2_RESNET50_CHANNELWISE_MODEL_ARCHIVE})
    endif()
    set(QUANT2_RESNET50_MODEL
        ${QUANT2_RESNET50_CHANNELWISE_MODEL_DIR}/ResNet50_qat_channelwise)
    inference_analysis_api_quant_test_run(
      test_analyzer_quant2_resnet50_channelwise_mkldnn
      ${QUANT_IMG_CLASS_TEST_APP} ${QUANT2_RESNET50_MODEL}
      ${QUANT2_RESNET50_MODEL} ${IMAGENET_DATA_PATH} true)

    ### Other tests

    # MKLDNN quantizer config
    set(MKLDNN_QUANTIZER_CONFIG_TEST_APP "test_mkldnn_quantizer_config")
    set(MKLDNN_QUANTIZER_CONFIG_TEST_APP_SRC
        "mkldnn_quantizer_config_tester.cc")
    inference_analysis_api_test_build(${MKLDNN_QUANTIZER_CONFIG_TEST_APP}
                                      ${MKLDNN_QUANTIZER_CONFIG_TEST_APP_SRC})
    inference_analysis_test_run(test_mkldnn_quantizer_config COMMAND
                                ${MKLDNN_QUANTIZER_CONFIG_TEST_APP})

    # preprocess data2bin imagenet
    download_int8_data_without_verify(${INT8_DATA_DIR} "imagenet_small.tar.gz")
    set(IMAGENET_SMALL_DATA_DIR "${INT8_DATA_DIR}/imagenet_small")
    set(IMAGENET_SMALL_OUTPUT_FILE "imagenet_small.bin")
    preprocess_data2bin_test_run(
      preprocess_local_imagenet "full_ILSVRC2012_val_preprocess.py"
      ${IMAGENET_SMALL_DATA_DIR} ${IMAGENET_SMALL_OUTPUT_FILE})

    # preprocess data2bin pascalvoc
    download_int8_data_without_verify(${INT8_DATA_DIR} "pascalvoc_small.tar.gz")
    set(PASCALVOC_SMALL_DATA_DIR "${INT8_DATA_DIR}/pascalvoc_small")
    set(PASCALVOC_SMALL_OUTPUT_FILE "pascalvoc_small.bin")
    preprocess_data2bin_test_run(
      preprocess_local_pascalvoc "full_pascalvoc_test_preprocess.py"
      ${PASCALVOC_SMALL_DATA_DIR} ${PASCALVOC_SMALL_OUTPUT_FILE})

  endif()

  # bert, max_len=20, embedding_dim=128
  set(BERT_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/bert_emb128")
  download_model_and_data_without_verify(
    ${BERT_INSTALL_DIR} "bert_emb128_model.tar.gz" "bert_data_len20.txt.tar.gz")
  inference_analysis_api_test(test_analyzer_bert ${BERT_INSTALL_DIR}
                              analyzer_bert_tester.cc)

  # multiple models prediction
  set(MMP_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/multi_model_prediction")
  download_data_without_verify(${MMP_INSTALL_DIR}
                               PaddleInference/mobilenet_v2_models.tar.gz)
  inference_multiple_models_analysis_api_test(
    test_analyzer_multi_model_prediction ${MMP_INSTALL_DIR}
    analyzer_mmp_tester.cc)

  if(WITH_GPU AND TENSORRT_FOUND)
    set(TRT_MODEL_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/trt_models")
    if(NOT EXISTS ${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models.tar.gz)
      inference_download_and_uncompress(
        ${TRT_MODEL_INSTALL_DIR} ${INFERENCE_URL}/tensorrt_test
        "trt_inference_test_models.tar.gz" 3dcccdc38b549b6b1b4089723757bd98)
    endif()
    set(TEST_SPLIT_CONVERTER_MODEL
        "${TRT_MODEL_INSTALL_DIR}/trt_split_op_converter_test")
    if(NOT EXISTS ${TEST_SPLIT_CONVERTER_MODEL}/split_converter.tgz)
      inference_download_and_uncompress_without_verify(
        ${TEST_SPLIT_CONVERTER_MODEL} ${INFERENCE_URL}/tensorrt_test
        "split_converter.tgz")
    endif()
    inference_analysis_test(
      trt_mobilenet_test
      SRCS
      trt_mobilenet_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    inference_analysis_test(
      trt_resnet50_test
      SRCS
      trt_resnet50_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    inference_analysis_test(
      trt_resnext_test
      SRCS
      trt_resnext_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    inference_analysis_test(
      trt_fc_prelu_test
      SRCS
      trt_fc_prelu_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    inference_analysis_test(
      trt_cascade_rcnn_test
      SRCS
      trt_cascade_rcnn_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    inference_analysis_test(
      trt_split_converter_test
      SRCS
      trt_split_converter_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TEST_SPLIT_CONVERTER_MODEL}/)
    inference_analysis_test(
      test_analyzer_capi_exp_gpu
      SRCS
      analyzer_capi_exp_gpu_tester.cc
      EXTRA_DEPS
      ${INFERENCE_C_EXTRA_DEPS}
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    if(WIN32)
      target_link_libraries(test_analyzer_capi_exp_gpu
                            paddle_inference_c_shared)
    else()
      target_link_libraries(test_analyzer_capi_exp_gpu paddle_inference_c)
    endif()
    inference_analysis_test(
      test_analyzer_capi_exp_xpu
      SRCS
      analyzer_capi_exp_xpu_tester.cc
      EXTRA_DEPS
      ${INFERENCE_C_EXTRA_DEPS}
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    if(WIN32)
      target_link_libraries(test_analyzer_capi_exp_xpu
                            paddle_inference_c_shared)
    else()
      target_link_libraries(test_analyzer_capi_exp_xpu paddle_inference_c)
    endif()

    #TODO(inference): windows encounter a SEH error, we need to fix it.
    if(NOT WIN32)
      inference_analysis_test(
        trt_rebind_stream_test
        SRCS
        trt_rebind_stream_test.cc
        EXTRA_DEPS
        paddle_inference_shared
        ARGS
        --infer_model=${TRT_MODEL_INSTALL_DIR}/trt_inference_test_models)
    endif()

    set(TRT_MODEL_QUANT_RESNET_DIR
        "${INFERENCE_DEMO_INSTALL_DIR}/small_quant_model")
    if(NOT EXISTS ${INFERENCE_DEMO_INSTALL_DIR}/small_quant_model.tgz)
      inference_download_and_uncompress_without_verify(
        ${INFERENCE_DEMO_INSTALL_DIR} ${INFERENCE_URL}/tensorrt_test
        "small_quant_model.tgz")
    endif()
    inference_analysis_test(
      trt_quant_int8_test
      SRCS
      trt_quant_int8_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_QUANT_RESNET_DIR})

    set(TRT_MODEL_QUANT_YOLOV3_DIR
        "${INFERENCE_DEMO_INSTALL_DIR}/yolov3_r50_quant_aware")
    if(NOT EXISTS ${INFERENCE_DEMO_INSTALL_DIR}/yolov3_r50_quant_aware.tgz)
      inference_download_and_uncompress_without_verify(
        ${INFERENCE_DEMO_INSTALL_DIR} ${INFERENCE_URL}/tensorrt_test
        "yolov3_r50_quant_aware.tgz")
    endif()
    inference_analysis_test(
      trt_quant_int8_yolov3_r50_test
      SRCS
      trt_quant_int8_yolov3_r50_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_QUANT_YOLOV3_DIR})

    set(TEST_TRT_DYNAMIC_MODEL2
        "${TRT_MODEL_INSTALL_DIR}/complex_model_dynamic")
    if(NOT EXISTS ${TEST_TRT_DYNAMIC_MODEL2}/complex_model_dynamic2.tar.gz)
      inference_download_and_uncompress_without_verify(
        ${TEST_TRT_DYNAMIC_MODEL2} ${INFERENCE_URL}/tensorrt_test
        "complex_model_dynamic2.tar.gz")
    endif()

    set(TEST_TRT_DYNAMIC_MODEL
        "${TRT_MODEL_INSTALL_DIR}/conv_bn_swish_split_gelu")
    if(NOT EXISTS ${TEST_TRT_DYNAMIC_MODEL}/conv_bn_swish_split_gelu.tar.gz)
      inference_download_and_uncompress(
        ${TEST_TRT_DYNAMIC_MODEL} ${INFERENCE_URL}/tensorrt_test
        "conv_bn_swish_split_gelu.tar.gz" 2a5e8791e47b221b4f782151d76da9c6)
    endif()
    inference_analysis_test(
      trt_dynamic_shape_test
      SRCS
      trt_dynamic_shape_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TRT_MODEL_INSTALL_DIR})

    set(TEST_TRT_ERNIE_MODEL "${TRT_MODEL_INSTALL_DIR}/ernie_test")
    if(NOT EXISTS ${TEST_TRT_ERNIE_MODEL}/ernie_model_4.tar.gz)
      inference_download_and_uncompress(
        ${TEST_TRT_ERNIE_MODEL} ${INFERENCE_URL}/tensorrt_test
        "ernie_model_4.tar.gz" 5fa371efa75706becbaad79195d2ca68)
    endif()

    inference_analysis_test(
      test_trt_dynamic_shape_ernie
      SRCS
      trt_dynamic_shape_ernie_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TEST_TRT_ERNIE_MODEL}/ernie_model_4)

    set(TEST_TRT_TRANSFORMER_PRUNE_MODEL
        "${TRT_MODEL_INSTALL_DIR}/transformer_prune")
    if(NOT EXISTS ${TEST_TRT_TRANSFORMER_PRUNE_MODEL}/transformer_prune.tar.gz)
      inference_download_and_uncompress(
        ${TEST_TRT_TRANSFORMER_PRUNE_MODEL} ${INFERENCE_URL}/tensorrt_test
        "transformer_prune.tar.gz" 77b56dc73ff0cf44ddb1ce9ca0b0f471)
    endif()

    inference_analysis_test(
      test_trt_dynamic_shape_transformer_prune
      SRCS
      trt_dynamic_shape_transformer_prune_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TEST_TRT_TRANSFORMER_PRUNE_MODEL}/transformer_prune)

    if(NOT EXISTS ${TEST_TRT_ERNIE_MODEL}/ernie_model_4_unserialized.tgz)
      inference_download_and_uncompress(
        ${TEST_TRT_ERNIE_MODEL} ${INFERENCE_URL}/tensorrt_test
        "ernie_model_4_unserialized.tgz" 833d73fc6a7f7e1ee4a1fd6419209e55)
    endif()

    inference_analysis_test(
      test_trt_dynamic_shape_ernie_ser_deser
      SRCS
      trt_dynamic_shape_ernie_serialize_deserialize_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TEST_TRT_ERNIE_MODEL}/ernie_model_4_unserialized)

    if(NOT EXISTS ${TEST_TRT_ERNIE_MODEL}/ernie_model_4_fp16_unserialized.tgz)
      inference_download_and_uncompress(
        ${TEST_TRT_ERNIE_MODEL} ${INFERENCE_URL}/tensorrt_test
        "ernie_model_4_fp16_unserialized.tgz" c5ff2d0cad79953ffbf2b8b9e2fae6e4)
    endif()

    inference_analysis_test(
      test_trt_dynamic_shape_ernie_fp16_ser_deser
      SRCS
      trt_dynamic_shape_ernie_fp16_serialize_deserialize_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${TEST_TRT_ERNIE_MODEL}/ernie_model_4_fp16_unserialized)

  endif()

  set(LITE_MODEL_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/lite")
  download_data_without_verify(${LITE_MODEL_INSTALL_DIR} "mul_model_fp32.tgz")

  inference_analysis_test(
    lite_mul_model_test
    SRCS
    lite_mul_model_test.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${LITE_MODEL_INSTALL_DIR})
  inference_analysis_test(
    lite_resnet50_test
    SRCS
    lite_resnet50_test.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${RESNET50_MODEL_DIR})

  inference_analysis_test(
    test_analyzer_capi_exp
    SRCS
    analyzer_capi_exp_tester.cc
    EXTRA_DEPS
    ${INFERENCE_C_EXTRA_DEPS}
    python
    ARGS
    --infer_model=${RESNET50_MODEL_DIR}/model)
  if(WIN32)
    target_link_libraries(test_analyzer_capi_exp paddle_inference_c_shared)
  else()
    target_link_libraries(test_analyzer_capi_exp paddle_inference_c)
  endif()

  inference_analysis_test(
    test_analyzer_capi_exp_pd_config
    SRCS
    analyzer_capi_exp_pd_config_tester.cc
    EXTRA_DEPS
    ${INFERENCE_C_EXTRA_DEPS}
    python
    ARGS
    --infer_model=${MOBILENET_INSTALL_DIR}/model)
  if(WIN32)
    target_link_libraries(test_analyzer_capi_exp_pd_config
                          paddle_inference_c_shared)
  else()
    target_link_libraries(test_analyzer_capi_exp_pd_config paddle_inference_c)
  endif()

  inference_analysis_test(
    test_analyzer_capi_exp_pd_tensor
    SRCS
    analyzer_capi_exp_pd_tensor_tester.cc
    EXTRA_DEPS
    ${INFERENCE_C_EXTRA_DEPS}
    python
    ARGS
    --infer_model=${MOBILENET_INSTALL_DIR}/model)
  if(WIN32)
    target_link_libraries(test_analyzer_capi_exp_pd_tensor
                          paddle_inference_c_shared)
  else()
    target_link_libraries(test_analyzer_capi_exp_pd_tensor paddle_inference_c)
  endif()

  if(NOT APPLE AND NOT WIN32)
    inference_analysis_test(
      test_analyzer_capi_exp_pd_threads
      SRCS
      analyzer_capi_exp_pd_threads_tester.cc
      EXTRA_DEPS
      ${INFERENCE_C_EXTRA_DEPS}
      python
      ARGS
      --infer_model=${MOBILENET_INSTALL_DIR}/model)
    if(WIN32)
      target_link_libraries(test_analyzer_capi_exp_pd_threads
                            paddle_inference_c_shared)
    else()
      target_link_libraries(test_analyzer_capi_exp_pd_threads
                            paddle_inference_c)
    endif()
  endif()

  inference_analysis_test(
    test_analyzer_zerocopytensor_tensor
    SRCS
    analyzer_zerocopy_tensor_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${OCR_INSTALL_DIR}/model)

  if(WITH_DISTRIBUTE AND WITH_PSCORE)
    inference_analysis_test(
      test_analyzer_dist_model
      SRCS
      analyzer_dist_model_tester.cc
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${OCR_INSTALL_DIR}/model)
  endif()

  if(WITH_DISTRIBUTE
     AND WITH_PSCORE
     AND WITH_XPU
     AND WITH_XPU_BKCL)
    inference_analysis_test(
      test_analyzer_dist_model_xpu
      SRCS
      analyzer_dist_model_xpu_tester.cc
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${OCR_INSTALL_DIR}/model)
  endif()

  inference_analysis_test(
    test_analyzer_paddletensor_tensor
    SRCS
    analyzer_paddle_tensor_tester.cc
    EXTRA_DEPS
    paddle_inference_shared
    python
    ARGS
    --infer_model=${OCR_INSTALL_DIR}/model
    --infer_data=${OCR_INSTALL_DIR}/data.txt
    --refer_result=${OCR_INSTALL_DIR}/result.txt)

  if(WITH_MKLDNN)
    inference_analysis_test(
      test_analyzer_capi_exp_int
      SRCS
      analyzer_capi_exp_int_tester.cc
      EXTRA_DEPS
      ${INFERENCE_C_EXTRA_DEPS}
      ARGS
      --infer_model=${INT8_DATA_DIR}/resnet50/model)
    if(WIN32)
      target_link_libraries(test_analyzer_capi_exp_int
                            paddle_inference_c_shared)
    else()
      target_link_libraries(test_analyzer_capi_exp_int paddle_inference_c)
    endif()
  endif()

  inference_analysis_test(
    test_analyzer_capi_exp_ner
    SRCS
    analyzer_capi_exp_ner_tester.cc
    EXTRA_DEPS
    ${INFERENCE_C_EXTRA_DEPS}
    python
    ARGS
    --infer_model=${CHINESE_NER_INSTALL_DIR}/model)
  if(WIN32)
    target_link_libraries(test_analyzer_capi_exp_ner paddle_inference_c_shared)
  else()
    target_link_libraries(test_analyzer_capi_exp_ner paddle_inference_c)
  endif()

  if(WITH_GPU)
    inference_analysis_test(
      paddle_infer_api_test
      SRCS
      paddle_infer_api_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${RESNET50_MODEL_DIR})

    inference_analysis_test(
      paddle_infer_api_copy_tensor_tester
      SRCS
      paddle_infer_api_copy_tensor_tester.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${RESNET50_MODEL_DIR})
    set_tests_properties(paddle_infer_api_copy_tensor_tester PROPERTIES TIMEOUT
                                                                        30)
  endif()

  cc_test(
    paddle_infer_api_errors_test
    SRCS paddle_infer_api_errors_tester.cc
    DEPS ${inference_api_tester_deps})

  if(WITH_GPU AND TENSORRT_FOUND)
    set_tests_properties(trt_resnext_test PROPERTIES TIMEOUT 300)
    set_tests_properties(trt_quant_int8_yolov3_r50_test PROPERTIES TIMEOUT 400)
    set_tests_properties(trt_resnet50_test PROPERTIES TIMEOUT 300)
    set_tests_properties(trt_cascade_rcnn_test PROPERTIES TIMEOUT 300)
    set_tests_properties(test_trt_dynamic_shape_ernie_ser_deser
                         PROPERTIES TIMEOUT 300)
    set_tests_properties(test_trt_dynamic_shape_ernie_fp16_ser_deser
                         PROPERTIES TIMEOUT 300)
    set_tests_properties(test_trt_dynamic_shape_ernie PROPERTIES TIMEOUT 480)
  endif()

  if(WITH_MKLDNN)
    set_tests_properties(test_analyzer_int8_resnet50 PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_int8_mobilenet_ssd PROPERTIES TIMEOUT
                                                                     120)
    set_tests_properties(test_analyzer_quant_performance_benchmark
                         PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_int8_mobilenetv2 PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_int8_mobilenetv1 PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_int8_mobilenetv3_large PROPERTIES TIMEOUT
                                                                         120)
    set_tests_properties(test_analyzer_quant2_mobilenetv1_mkldnn
                         PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_quant2_resnet50_channelwise_mkldnn
                         PROPERTIES TIMEOUT 120)
  endif()

  set_tests_properties(lite_resnet50_test PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_mobilenet_transpose PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_resnet50 PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_ner PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_ernie_int8 PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_googlenet PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_small_dam PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_transformer PROPERTIES TIMEOUT 120)
  set_tests_properties(test_analyzer_mobilenet_depthwise_conv PROPERTIES TIMEOUT
                                                                         120)
  if(WITH_GPU)
    set_tests_properties(test_analyzer_bert PROPERTIES TIMEOUT 120)
    set_tests_properties(test_analyzer_ernie PROPERTIES TIMEOUT 120)
  endif()
  if(WITH_GPU AND TENSORRT_FOUND)
    set_tests_properties(trt_mobilenet_test PROPERTIES TIMEOUT 240)
    if(NOT WIN32)
      set_tests_properties(trt_rebind_stream_test
                           PROPERTIES TIMEOUT 360 LABELS "RUN_TYPE=EXCLUSIVE")
    endif()
    if(WITH_MKLDNN)
      set_tests_properties(test_analyzer_bfloat16_resnet50 PROPERTIES TIMEOUT
                                                                      120)
    endif()
  endif()
  if(ON_INFER OR WITH_GPU)
    set_tests_properties(test_analyzer_transformer_profile PROPERTIES TIMEOUT
                                                                      120)
  endif()

  if(WITH_IPU)
    #word2vec sample
    set(WORD2VEC_INSTALL_DIR
        "${INFERENCE_DEMO_INSTALL_DIR}/word2vec/word2vec.inference.model")
    inference_analysis_test(
      ipu_word2vec_sample
      SRCS
      ipu_word2vec_sample.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${WORD2VEC_INSTALL_DIR})

    # ERNIE
    set(ERNIE_INSTALL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/Ernie")
    inference_analysis_api_test(
      ipu_ernie_test ${ERNIE_INSTALL_DIR} ipu_ernie_test.cc ARGS --warmup=true
      --repeat=10)
    inference_analysis_api_test(
      ipu_ernie_fp16_test ${ERNIE_INSTALL_DIR} ipu_ernie_fp16_test.cc ARGS
      --warmup=true --repeat=10)

    # Resnet50
    set(RESNET50_MODEL_DIR "${INFERENCE_DEMO_INSTALL_DIR}/resnet50")
    inference_analysis_test(
      ipu_resnet50_test
      SRCS
      ipu_resnet50_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${RESNET50_MODEL_DIR}
      --warmup=true
      --repeat=10)
    inference_analysis_test(
      ipu_resnet50_fp16_test
      SRCS
      ipu_resnet50_fp16_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      ARGS
      --infer_model=${RESNET50_MODEL_DIR}
      --warmup=true
      --repeat=10)

    # Only support Resnet50 and Ernie currently
    inference_analysis_api_test(
      ipu_multi_model_profile
      SRCS
      ipu_multi_model_profile.cc
      ARGS
      --model_name="Resnet50"
      --infer_model=${RESNET50_MODEL_DIR}
      --warmup=true
      --repeat=10)
  endif()

  if(WITH_XPU)
    inference_analysis_test(
      xpu_config_resnet50_test
      SRCS
      xpu_config_resnet50_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${RESNET50_MODEL_DIR})
    inference_analysis_test(
      xpu_runtime_config_resnet50_test
      SRCS
      xpu_runtime_config_resnet50_test.cc
      EXTRA_DEPS
      paddle_inference_shared
      python
      ARGS
      --infer_model=${RESNET50_MODEL_DIR})
  endif()

  set(inference_deps ${analysis_deps} ${inference_api_tester_deps} analysis
                     naive_executor ${GLOB_PASS_LIB})

  if(WITH_TESTING)
    if(NOT APPLE AND NOT WIN32)
      inference_base_test(
        test_api_impl
        SRCS
        api_impl_tester.cc
        DEPS
        paddle_inference_shared
        python
        ARGS
        --word2vec_dirname=${WORD2VEC_MODEL_DIR}
        --book_dirname=${IMG_CLS_RESNET_INSTALL_DIR})
    elseif(WIN32)
      inference_base_test(
        test_api_impl
        SRCS
        api_impl_tester.cc
        DEPS
        ${inference_deps}
        ARGS
        --word2vec_dirname=${WORD2VEC_MODEL_DIR}
        --book_dirname=${IMG_CLS_RESNET_INSTALL_DIR})
    endif()
  endif()

  if(NOT APPLE AND NOT WIN32)
    cc_test_old(
      test_analysis_predictor
      SRCS
      analysis_predictor_tester.cc
      DEPS
      paddle_inference_shared
      python
      ARGS
      --dirname=${WORD2VEC_MODEL_DIR})
  elseif(WIN32)
    cc_test_old(
      test_analysis_predictor
      SRCS
      analysis_predictor_tester.cc
      DEPS
      analysis_predictor
      benchmark
      ${inference_deps}
      ARGS
      --dirname=${WORD2VEC_MODEL_DIR})
  endif()

  if(WITH_TESTING AND WITH_MKLDNN)
    if(NOT APPLE AND NOT WIN32)
      cc_test(
        test_mkldnn_quantizer
        SRCS mkldnn_quantizer_tester.cc
        DEPS paddle_inference_shared python ARGS
             --dirname=${WORD2VEC_MODEL_DIR})
    elseif(WIN32)
      cc_test(
        test_mkldnn_quantizer
        SRCS mkldnn_quantizer_tester.cc
        DEPS analysis_predictor benchmark ${inference_deps} ARGS
             --dirname=${WORD2VEC_MODEL_DIR})
    endif()
  endif()

  if(WITH_TESTING AND TEST test_api_impl)
    if(NOT APPLE)
      set_tests_properties(test_api_impl PROPERTIES TIMEOUT 120)
    endif()
  endif()
endif()
