未验证 提交 042c3515 编写于 作者: littletomatodonkey's avatar littletomatodonkey 提交者: GitHub

fix mkldnn thread num (#1041)

上级 5cd7a05d
...@@ -2,8 +2,8 @@ Global: ...@@ -2,8 +2,8 @@ Global:
rec_inference_model_dir: "./models/cartoon_rec_ResNet50_iCartoon_v1.0_infer/" rec_inference_model_dir: "./models/cartoon_rec_ResNet50_iCartoon_v1.0_infer/"
batch_size: 1 batch_size: 1
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -2,8 +2,8 @@ Global: ...@@ -2,8 +2,8 @@ Global:
rec_inference_model_dir: "./models/logo_rec_ResNet50_Logo3K_v1.0_infer/" rec_inference_model_dir: "./models/logo_rec_ResNet50_Logo3K_v1.0_infer/"
batch_size: 1 batch_size: 1
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -2,8 +2,8 @@ Global: ...@@ -2,8 +2,8 @@ Global:
rec_inference_model_dir: "./models/product_ResNet50_vd_aliproduct_v1.0_infer" rec_inference_model_dir: "./models/product_ResNet50_vd_aliproduct_v1.0_infer"
batch_size: 1 batch_size: 1
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -2,8 +2,8 @@ Global: ...@@ -2,8 +2,8 @@ Global:
rec_inference_model_dir: "./models/vehicle_cls_ResNet50_CompCars_v1.0_infer/" rec_inference_model_dir: "./models/vehicle_cls_ResNet50_CompCars_v1.0_infer/"
batch_size: 1 batch_size: 1
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -12,8 +12,8 @@ Global: ...@@ -12,8 +12,8 @@ Global:
- foreground - foreground
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -3,8 +3,8 @@ Global: ...@@ -3,8 +3,8 @@ Global:
inference_model_dir: "./models" inference_model_dir: "./models"
batch_size: 1 batch_size: 1
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
...@@ -29,4 +29,4 @@ PostProcess: ...@@ -29,4 +29,4 @@ PostProcess:
topk: 5 topk: 5
class_id_map_file: "../ppcls/utils/imagenet1k_label_list.txt" class_id_map_file: "../ppcls/utils/imagenet1k_label_list.txt"
SavePreLabel: SavePreLabel:
save_dir: ./pre_label/ save_dir: ./pre_label/
\ No newline at end of file
...@@ -10,8 +10,8 @@ Global: ...@@ -10,8 +10,8 @@ Global:
# inference engine config # inference engine config
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -13,8 +13,8 @@ Global: ...@@ -13,8 +13,8 @@ Global:
# inference engine config # inference engine config
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -13,8 +13,8 @@ Global: ...@@ -13,8 +13,8 @@ Global:
# inference engine config # inference engine config
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -10,8 +10,8 @@ Global: ...@@ -10,8 +10,8 @@ Global:
# inference engine config # inference engine config
use_gpu: False use_gpu: False
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
...@@ -13,8 +13,8 @@ Global: ...@@ -13,8 +13,8 @@ Global:
# inference engine config # inference engine config
use_gpu: True use_gpu: True
enable_mkldnn: False enable_mkldnn: True
cpu_num_threads: 100 cpu_num_threads: 10
enable_benchmark: True enable_benchmark: True
use_fp16: False use_fp16: False
ir_optim: True ir_optim: True
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册