未验证 提交 d01f92a6 编写于 作者: C cuicheng01 提交者: GitHub

Merge pull request #1858 from cuicheng01/update_benchmark

Update benchmark config
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:32
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:32
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:32
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:8
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:64
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran ...@@ -50,9 +50,3 @@ inference:python/predict_cls.py -c configs/inference_cls.yaml -o PreProcess.tran
-o Global.benchmark:True -o Global.benchmark:True
null:null null:null
null:null null:null
===========================train_benchmark_params==========================
batch_size:128
fp_items:fp32
epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:FLAGS_eager_delete_tensor_gb=0.0;FLAGS_fraction_of_gpu_memory_to_use=0.98;FLAGS_conv_workspace_size_limit=4096
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册