未验证 提交 de9b3d82 编写于 作者: S shangliang Xu 提交者: GitHub

[TIPC] add inference benchmark param (#5301)

上级 2f9e086a
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,1024,1024]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,1024,1024]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,1024,1024]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,256,192]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,256,192]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,512,512]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,256,192]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/keypoint_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,128,96]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
--trt_max_shape:1600 --trt_max_shape:1600
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,1088]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:False --run_benchmark:False
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,576]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:False --run_benchmark:False
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,576]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,1088]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/pptracking/python/mot_jde_infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:False --run_benchmark:False
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,576]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,640,640]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,416,416]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,416,416]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,416,416]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,640,640]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,416,416]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,512,512]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,320,320]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,640,640]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,640,640]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:False --run_benchmark:False
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,512,864]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,800,1344]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,300,300]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,512,512]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -55,3 +55,5 @@ fp_items:fp32|fp16 ...@@ -55,3 +55,5 @@ fp_items:fp32|fp16
epoch:1 epoch:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py ...@@ -49,3 +49,5 @@ inference:./deploy/python/infer.py
--save_log_path:null --save_log_path:null
--run_benchmark:True --run_benchmark:True
null:null null:null
===========================infer_benchmark_params===========================
random_infer_input:[{float32,[3,608,608]},{float32,[2]},{float32,[2]}]
\ No newline at end of file
...@@ -263,7 +263,7 @@ else ...@@ -263,7 +263,7 @@ else
fi fi
if [ ${autocast} = "amp" ]; then if [ ${autocast} = "amp" ]; then
set_autocast="--fp16" set_autocast="--amp"
else else
set_autocast=" " set_autocast=" "
fi fi
...@@ -299,7 +299,7 @@ else ...@@ -299,7 +299,7 @@ else
if [ ${run_export} != "null" ]; then if [ ${run_export} != "null" ]; then
# run export model # run export model
set_export_weight=$(func_set_params "${export_weight_key}" "${save_log}/${model_name}/${train_model_name}") set_export_weight=$(func_set_params "${export_weight_key}" "${save_log}/${model_name}/${train_model_name}")
set_save_export_dir=$(func_set_params "${save_export_key}" "${save_export_value}") set_save_export_dir=$(func_set_params "${save_export_key}" "${save_log}")
export_cmd="${python} ${run_export} ${set_export_weight} ${set_filename} ${set_save_export_dir} " export_cmd="${python} ${run_export} ${set_export_weight} ${set_filename} ${set_save_export_dir} "
eval $export_cmd eval $export_cmd
status_check $? "${export_cmd}" "${status_log}" status_check $? "${export_cmd}" "${status_log}"
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册