提交 68f636d7 编写于 作者: H HydrogenSulfate

simplify train_infer_python.txt configs

上级 f729688e
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=288 -o PreProcess.transform_ops.1.CropImage.size=256 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=288 -o PreProcess.transform_ops.1.CropImage.size=256
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=272 -o PreProcess.transform_ops.1.CropImage.size=240 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=272 -o PreProcess.transform_ops.1.CropImage.size=240
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=260 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=260
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=332 -o PreProcess.transform_ops.1.CropImage.size=300 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=332 -o PreProcess.transform_ops.1.CropImage.size=300
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=412 -o PreProcess.transform_ops.1.CropImage.size=380 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=412 -o PreProcess.transform_ops.1.CropImage.size=380
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=488 -o PreProcess.transform_ops.1.CropImage.size=456 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=488 -o PreProcess.transform_ops.1.CropImage.size=456
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=560 -o PreProcess.transform_ops.1.CropImage.size=528 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=560 -o PreProcess.transform_ops.1.CropImage.size=528
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=632 -o PreProcess.transform_ops.1.CropImage.size=600 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=632 -o PreProcess.transform_ops.1.CropImage.size=600
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_rec.py -c configs/inference_rec.yaml inference:python/predict_rec.py -c configs/inference_rec.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.rec_inference_model_dir:../inference -o Global.rec_inference_model_dir:../inference
-o Global.infer_imgs:../dataset/Aliproduct/demo_test/ -o Global.infer_imgs:../dataset/Aliproduct/demo_test/
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=236 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=236
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=232 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=232
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
此差异已折叠。
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册