提交 68f636d7 编写于 作者: H HydrogenSulfate

simplify train_infer_python.txt configs

上级 f729688e
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=288 -o PreProcess.transform_ops.1.CropImage.size=256 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=288 -o PreProcess.transform_ops.1.CropImage.size=256
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=272 -o PreProcess.transform_ops.1.CropImage.size=240 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=272 -o PreProcess.transform_ops.1.CropImage.size=240
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=260 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=260
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=332 -o PreProcess.transform_ops.1.CropImage.size=300 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=332 -o PreProcess.transform_ops.1.CropImage.size=300
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=412 -o PreProcess.transform_ops.1.CropImage.size=380 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=412 -o PreProcess.transform_ops.1.CropImage.size=380
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=488 -o PreProcess.transform_ops.1.CropImage.size=456 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=488 -o PreProcess.transform_ops.1.CropImage.size=456
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=560 -o PreProcess.transform_ops.1.CropImage.size=528 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=560 -o PreProcess.transform_ops.1.CropImage.size=528
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=632 -o PreProcess.transform_ops.1.CropImage.size=600 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=632 -o PreProcess.transform_ops.1.CropImage.size=600
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_rec.py -c configs/inference_rec.yaml inference:python/predict_rec.py -c configs/inference_rec.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.rec_inference_model_dir:../inference -o Global.rec_inference_model_dir:../inference
-o Global.infer_imgs:../dataset/Aliproduct/demo_test/ -o Global.infer_imgs:../dataset/Aliproduct/demo_test/
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=292 -o PreProcess.transform_ops.1.CropImage.size=256 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.,0.,0.] -o PreProcess.transform_ops.2.NormalizeImage.std=[1.,1.,1.]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=236 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=236
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=232 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=232
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
\ No newline at end of file
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
random_infer_input:[{float32,[3,224,224]}] random_infer_input:[{float32,[3,224,224]}]
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=256 -o PreProcess.transform_ops.1.CropImage.size=224
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================train_benchmark_params========================== ===========================train_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.5,0.5,0.5] -o PreProcess.transform_ops.2.NormalizeImage.std=[0.5,0.5,0.5] inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=248 -o PreProcess.transform_ops.2.NormalizeImage.mean=[0.5,0.5,0.5] -o PreProcess.transform_ops.2.NormalizeImage.std=[0.5,0.5,0.5]
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=384 -o PreProcess.transform_ops.1.CropImage.size=384
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml inference:python/predict_cls.py -c configs/inference_cls.yaml
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -39,15 +39,15 @@ infer_export:True ...@@ -39,15 +39,15 @@ infer_export:True
infer_quant:Fasle infer_quant:Fasle
inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299 inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.transform_ops.0.ResizeImage.resize_short=320 -o PreProcess.transform_ops.1.CropImage.size=299
-o Global.use_gpu:True|False -o Global.use_gpu:True|False
-o Global.enable_mkldnn:True|False -o Global.enable_mkldnn:False
-o Global.cpu_num_threads:1|6 -o Global.cpu_num_threads:1
-o Global.batch_size:1|16 -o Global.batch_size:1
-o Global.use_tensorrt:True|False -o Global.use_tensorrt:False
-o Global.use_fp16:True|False -o Global.use_fp16:False
-o Global.inference_model_dir:../inference -o Global.inference_model_dir:../inference
-o Global.infer_imgs:../dataset/ILSVRC2012/val -o Global.infer_imgs:../dataset/ILSVRC2012/val
-o Global.save_log_path:null -o Global.save_log_path:null
-o Global.benchmark:True -o Global.benchmark:False
null:null null:null
null:null null:null
===========================infer_benchmark_params========================== ===========================infer_benchmark_params==========================
......
...@@ -172,17 +172,32 @@ if [[ $FILENAME == *use_dali* ]]; then ...@@ -172,17 +172,32 @@ if [[ $FILENAME == *use_dali* ]]; then
fi fi
if [[ ${MODE} = "lite_train_lite_infer" ]] || [[ ${MODE} = "lite_train_whole_infer" ]]; then if [[ ${MODE} = "lite_train_lite_infer" ]] || [[ ${MODE} = "lite_train_whole_infer" ]]; then
# pretrain lite train data if [[ ${model_name} =~ "GeneralRecognition" ]]; then
cd dataset cd dataset
rm -rf ILSVRC2012 rm -rf Aliproduct
wget -nc https://paddle-imagenet-models-name.bj.bcebos.com/data/whole_chain/whole_chain_little_train.tar --no-check-certificate rm -rf train_reg_all_data.txt
tar xf whole_chain_little_train.tar rm -rf demo_train
ln -s whole_chain_little_train ILSVRC2012 wget -nc https://paddle-imagenet-models-name.bj.bcebos.com/data/whole_chain/tipc_shitu_demo_data.tar --no-check-certificate
cd ILSVRC2012 tar -xf tipc_shitu_demo_data.tar
mv train.txt train_list.txt ln -s tipc_shitu_demo_data Aliproduct
mv val.txt val_list.txt ln -s tipc_shitu_demo_data/demo_train.txt train_reg_all_data.txt
cp -r train/* val/ ln -s tipc_shitu_demo_data/demo_train demo_train
cd ../../ cd tipc_shitu_demo_data
ln -s demo_test.txt val_list.txt
cd ../../
else
# pretrain lite train data
cd dataset
rm -rf ILSVRC2012
wget -nc https://paddle-imagenet-models-name.bj.bcebos.com/data/whole_chain/whole_chain_little_train.tar --no-check-certificate
tar xf whole_chain_little_train.tar
ln -s whole_chain_little_train ILSVRC2012
cd ILSVRC2012
mv train.txt train_list.txt
mv val.txt val_list.txt
cp -r train/* val/
cd ../../
fi
if [[ ${FILENAME} =~ "pact_infer" ]]; then if [[ ${FILENAME} =~ "pact_infer" ]]; then
# download pretrained model for PACT training # download pretrained model for PACT training
pretrpretrained_model_url=$(func_parser_value "${lines[35]}") pretrpretrained_model_url=$(func_parser_value "${lines[35]}")
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册