From 936ead600144f5be5fe594577958ea8ef2549b6c Mon Sep 17 00:00:00 2001 From: dongshuilong Date: Tue, 12 Oct 2021 11:02:19 +0000 Subject: [PATCH] change infer batch_size to 1|16 for whole_train --- tests/config/DarkNet53.txt | 2 +- tests/config/GhostNet_x0_5.txt | 2 +- tests/config/GhostNet_x1_0.txt | 2 +- tests/config/GhostNet_x1_3.txt | 2 +- tests/config/HRNet_W18_C.txt | 2 +- tests/config/LeViT_128S.txt | 2 +- tests/config/MobileNetV1.txt | 2 +- tests/config/MobileNetV1_x0_25.txt | 2 +- tests/config/MobileNetV1_x0_5.txt | 2 +- tests/config/MobileNetV1_x0_75.txt | 2 +- tests/config/MobileNetV2.txt | 2 +- tests/config/MobileNetV2_x0_25.txt | 2 +- tests/config/MobileNetV2_x0_5.txt | 2 +- tests/config/MobileNetV2_x0_75.txt | 2 +- tests/config/MobileNetV2_x1_5.txt | 2 +- tests/config/MobileNetV2_x2_0.txt | 2 +- tests/config/MobileNetV3_large_x0_35.txt | 2 +- tests/config/MobileNetV3_large_x0_5.txt | 2 +- tests/config/MobileNetV3_large_x0_75.txt | 2 +- tests/config/MobileNetV3_large_x1_0.txt | 2 +- tests/config/MobileNetV3_large_x1_25.txt | 2 +- tests/config/MobileNetV3_small_x0_35.txt | 2 +- tests/config/MobileNetV3_small_x0_5.txt | 2 +- tests/config/MobileNetV3_small_x0_75.txt | 2 +- tests/config/MobileNetV3_small_x1_0.txt | 2 +- tests/config/MobileNetV3_small_x1_25.txt | 2 +- tests/config/PPLCNet_x0_25.txt | 2 +- tests/config/PPLCNet_x0_35.txt | 2 +- tests/config/PPLCNet_x0_5.txt | 2 +- tests/config/PPLCNet_x1_0.txt | 2 +- tests/config/ResNeXt101_vd_64x4d.txt | 2 +- tests/config/ResNet50_vd.txt | 2 +- tests/config/ShuffleNetV2_x0_25.txt | 2 +- tests/config/ShuffleNetV2_x0_33.txt | 2 +- tests/config/ShuffleNetV2_x0_5.txt | 2 +- tests/config/ShuffleNetV2_x1_0.txt | 2 +- tests/config/SwinTransformer_tiny_patch4_window7_224.txt | 2 +- 37 files changed, 37 insertions(+), 37 deletions(-) diff --git a/tests/config/DarkNet53.txt b/tests/config/DarkNet53.txt index 589d6551..9b9b95c8 100644 --- a/tests/config/DarkNet53.txt +++ b/tests/config/DarkNet53.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/GhostNet_x0_5.txt b/tests/config/GhostNet_x0_5.txt index 9505f9ad..b752da46 100644 --- a/tests/config/GhostNet_x0_5.txt +++ b/tests/config/GhostNet_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/GhostNet_x1_0.txt b/tests/config/GhostNet_x1_0.txt index bcef875c..c927b2bf 100644 --- a/tests/config/GhostNet_x1_0.txt +++ b/tests/config/GhostNet_x1_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/GhostNet_x1_3.txt b/tests/config/GhostNet_x1_3.txt index 7501dec5..5766332f 100644 --- a/tests/config/GhostNet_x1_3.txt +++ b/tests/config/GhostNet_x1_3.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/HRNet_W18_C.txt b/tests/config/HRNet_W18_C.txt index 0609047b..c5334e38 100644 --- a/tests/config/HRNet_W18_C.txt +++ b/tests/config/HRNet_W18_C.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/LeViT_128S.txt b/tests/config/LeViT_128S.txt index 35feb2ff..6ddc0cd9 100644 --- a/tests/config/LeViT_128S.txt +++ b/tests/config/LeViT_128S.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|Fasle -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV1.txt b/tests/config/MobileNetV1.txt index 27890621..319f852d 100644 --- a/tests/config/MobileNetV1.txt +++ b/tests/config/MobileNetV1.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV1_x0_25.txt b/tests/config/MobileNetV1_x0_25.txt index 7b827175..a46a8e30 100644 --- a/tests/config/MobileNetV1_x0_25.txt +++ b/tests/config/MobileNetV1_x0_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV1_x0_5.txt b/tests/config/MobileNetV1_x0_5.txt index b7ba44df..bca197b7 100644 --- a/tests/config/MobileNetV1_x0_5.txt +++ b/tests/config/MobileNetV1_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV1_x0_75.txt b/tests/config/MobileNetV1_x0_75.txt index dbab4deb..ba81e35c 100644 --- a/tests/config/MobileNetV1_x0_75.txt +++ b/tests/config/MobileNetV1_x0_75.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2.txt b/tests/config/MobileNetV2.txt index 3904cb5d..d4cf3fc5 100644 --- a/tests/config/MobileNetV2.txt +++ b/tests/config/MobileNetV2.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2_x0_25.txt b/tests/config/MobileNetV2_x0_25.txt index 7fb93555..5368aef9 100644 --- a/tests/config/MobileNetV2_x0_25.txt +++ b/tests/config/MobileNetV2_x0_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2_x0_5.txt b/tests/config/MobileNetV2_x0_5.txt index 5a36396d..ba35c3af 100644 --- a/tests/config/MobileNetV2_x0_5.txt +++ b/tests/config/MobileNetV2_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2_x0_75.txt b/tests/config/MobileNetV2_x0_75.txt index cf67fefd..0f5a8ab8 100644 --- a/tests/config/MobileNetV2_x0_75.txt +++ b/tests/config/MobileNetV2_x0_75.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2_x1_5.txt b/tests/config/MobileNetV2_x1_5.txt index 657bb245..55da65b1 100644 --- a/tests/config/MobileNetV2_x1_5.txt +++ b/tests/config/MobileNetV2_x1_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV2_x2_0.txt b/tests/config/MobileNetV2_x2_0.txt index 9ef45808..cd32244f 100644 --- a/tests/config/MobileNetV2_x2_0.txt +++ b/tests/config/MobileNetV2_x2_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_large_x0_35.txt b/tests/config/MobileNetV3_large_x0_35.txt index e139c0ae..bc6e77c0 100644 --- a/tests/config/MobileNetV3_large_x0_35.txt +++ b/tests/config/MobileNetV3_large_x0_35.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_large_x0_5.txt b/tests/config/MobileNetV3_large_x0_5.txt index e1ee2c53..693423da 100644 --- a/tests/config/MobileNetV3_large_x0_5.txt +++ b/tests/config/MobileNetV3_large_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_large_x0_75.txt b/tests/config/MobileNetV3_large_x0_75.txt index 685c1f6f..24b35de5 100644 --- a/tests/config/MobileNetV3_large_x0_75.txt +++ b/tests/config/MobileNetV3_large_x0_75.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_large_x1_0.txt b/tests/config/MobileNetV3_large_x1_0.txt index f290db3e..873072f2 100644 --- a/tests/config/MobileNetV3_large_x1_0.txt +++ b/tests/config/MobileNetV3_large_x1_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_large_x1_25.txt b/tests/config/MobileNetV3_large_x1_25.txt index 7f357083..34057be5 100644 --- a/tests/config/MobileNetV3_large_x1_25.txt +++ b/tests/config/MobileNetV3_large_x1_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_small_x0_35.txt b/tests/config/MobileNetV3_small_x0_35.txt index 40fc5987..0f8b75e9 100644 --- a/tests/config/MobileNetV3_small_x0_35.txt +++ b/tests/config/MobileNetV3_small_x0_35.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_small_x0_5.txt b/tests/config/MobileNetV3_small_x0_5.txt index d444054f..0693c086 100644 --- a/tests/config/MobileNetV3_small_x0_5.txt +++ b/tests/config/MobileNetV3_small_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_small_x0_75.txt b/tests/config/MobileNetV3_small_x0_75.txt index 95cd5b65..ce424273 100644 --- a/tests/config/MobileNetV3_small_x0_75.txt +++ b/tests/config/MobileNetV3_small_x0_75.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_small_x1_0.txt b/tests/config/MobileNetV3_small_x1_0.txt index 7e3e7e86..a1f2b8b7 100644 --- a/tests/config/MobileNetV3_small_x1_0.txt +++ b/tests/config/MobileNetV3_small_x1_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/MobileNetV3_small_x1_25.txt b/tests/config/MobileNetV3_small_x1_25.txt index 617b49cb..49831b66 100644 --- a/tests/config/MobileNetV3_small_x1_25.txt +++ b/tests/config/MobileNetV3_small_x1_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/PPLCNet_x0_25.txt b/tests/config/PPLCNet_x0_25.txt index c3166ec1..0777cfa4 100644 --- a/tests/config/PPLCNet_x0_25.txt +++ b/tests/config/PPLCNet_x0_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/PPLCNet_x0_35.txt b/tests/config/PPLCNet_x0_35.txt index 347661c9..0ce679e4 100644 --- a/tests/config/PPLCNet_x0_35.txt +++ b/tests/config/PPLCNet_x0_35.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/PPLCNet_x0_5.txt b/tests/config/PPLCNet_x0_5.txt index b57d27dd..fbee882a 100644 --- a/tests/config/PPLCNet_x0_5.txt +++ b/tests/config/PPLCNet_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/PPLCNet_x1_0.txt b/tests/config/PPLCNet_x1_0.txt index be0c45ac..1e991164 100644 --- a/tests/config/PPLCNet_x1_0.txt +++ b/tests/config/PPLCNet_x1_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ResNeXt101_vd_64x4d.txt b/tests/config/ResNeXt101_vd_64x4d.txt index 90b79655..aab18d56 100644 --- a/tests/config/ResNeXt101_vd_64x4d.txt +++ b/tests/config/ResNeXt101_vd_64x4d.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ResNet50_vd.txt b/tests/config/ResNet50_vd.txt index a2c7ea25..9b8f27cc 100644 --- a/tests/config/ResNet50_vd.txt +++ b/tests/config/ResNet50_vd.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ShuffleNetV2_x0_25.txt b/tests/config/ShuffleNetV2_x0_25.txt index 9ad0ebd5..1c80e4f4 100644 --- a/tests/config/ShuffleNetV2_x0_25.txt +++ b/tests/config/ShuffleNetV2_x0_25.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ShuffleNetV2_x0_33.txt b/tests/config/ShuffleNetV2_x0_33.txt index 70dc11a7..34e813fb 100644 --- a/tests/config/ShuffleNetV2_x0_33.txt +++ b/tests/config/ShuffleNetV2_x0_33.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ShuffleNetV2_x0_5.txt b/tests/config/ShuffleNetV2_x0_5.txt index 84f1df60..b918a780 100644 --- a/tests/config/ShuffleNetV2_x0_5.txt +++ b/tests/config/ShuffleNetV2_x0_5.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/ShuffleNetV2_x1_0.txt b/tests/config/ShuffleNetV2_x1_0.txt index a208fa14..1055910a 100644 --- a/tests/config/ShuffleNetV2_x1_0.txt +++ b/tests/config/ShuffleNetV2_x1_0.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference diff --git a/tests/config/SwinTransformer_tiny_patch4_window7_224.txt b/tests/config/SwinTransformer_tiny_patch4_window7_224.txt index f2937b14..94acb150 100644 --- a/tests/config/SwinTransformer_tiny_patch4_window7_224.txt +++ b/tests/config/SwinTransformer_tiny_patch4_window7_224.txt @@ -41,7 +41,7 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o Global.use_gpu:True|False -o Global.enable_mkldnn:True|False -o Global.cpu_num_threads:1|6 --o Global.batch_size:1 +-o Global.batch_size:1|16 -o Global.use_tensorrt:True|False -o Global.use_fp16:True|False -o Global.inference_model_dir:../inference -- GitLab