未验证 提交 5f14efd5 编写于 作者: L LokeZhou 提交者: GitHub

[TIPC] * benchamrk use cocomini dataset (#7995)

* benchamrk set run 5m finish
* static train pop data['image_file']
上级 2b2db894
...@@ -487,6 +487,9 @@ class Trainer(object): ...@@ -487,6 +487,9 @@ class Trainer(object):
profiler.add_profiler_step(profiler_options) profiler.add_profiler_step(profiler_options)
self._compose_callback.on_step_begin(self.status) self._compose_callback.on_step_begin(self.status)
data['epoch_id'] = epoch_id data['epoch_id'] = epoch_id
if self.cfg.get('to_static',
False) and 'image_file' in data.keys():
data.pop('image_file')
if self.use_amp: if self.use_amp:
if isinstance( if isinstance(
......
...@@ -243,7 +243,7 @@ for batch_size in ${batch_size_list[*]}; do ...@@ -243,7 +243,7 @@ for batch_size in ${batch_size_list[*]}; do
--run_mode ${run_mode} \ --run_mode ${run_mode} \
--fp_item ${precision} \ --fp_item ${precision} \
--keyword ips: \ --keyword ips: \
--skip_steps 2 \ --skip_steps 4 \
--device_num ${device_num} \ --device_num ${device_num} \
--speed_unit images/s \ --speed_unit images/s \
--convergence_key loss: " --convergence_key loss: "
...@@ -279,7 +279,7 @@ for batch_size in ${batch_size_list[*]}; do ...@@ -279,7 +279,7 @@ for batch_size in ${batch_size_list[*]}; do
--run_mode ${run_mode} \ --run_mode ${run_mode} \
--fp_item ${precision} \ --fp_item ${precision} \
--keyword ips: \ --keyword ips: \
--skip_steps 2 \ --skip_steps 4 \
--device_num ${device_num} \ --device_num ${device_num} \
--speed_unit images/s \ --speed_unit images/s \
--convergence_key loss: " --convergence_key loss: "
......
...@@ -53,6 +53,6 @@ inference:./deploy/python/infer.py ...@@ -53,6 +53,6 @@ inference:./deploy/python/infer.py
batch_size:2 batch_size:2
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:2 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
\ No newline at end of file
...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py ...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py
batch_size:2 batch_size:2
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:2 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
......
...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py ...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py
batch_size:2|8 batch_size:2|8
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:3 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
......
...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py ...@@ -53,7 +53,7 @@ inference:./deploy/python/infer.py
batch_size:2|4 batch_size:2|4
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:2 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
......
...@@ -53,7 +53,7 @@ null:null ...@@ -53,7 +53,7 @@ null:null
batch_size:64 batch_size:64
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:25 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
......
...@@ -53,8 +53,10 @@ inference:./deploy/python/infer.py ...@@ -53,8 +53,10 @@ inference:./deploy/python/infer.py
batch_size:8 batch_size:8
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:12 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
numpy_infer_input:3x640x640_2.npy numpy_infer_input:3x640x640_2.npy
===========================to_static_train_benchmark_params===========================
to_static_train:--to_static
...@@ -53,7 +53,7 @@ null:null ...@@ -53,7 +53,7 @@ null:null
batch_size:8 batch_size:8
fp_items:fp32|fp16 fp_items:fp32|fp16
epoch:1 epoch:1
repeat:3 repeat:1
--profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile --profiler_options:batch_range=[10,20];state=GPU;tracer_option=Default;profile_path=model.profile
flags:null flags:null
===========================infer_benchmark_params=========================== ===========================infer_benchmark_params===========================
......
...@@ -99,9 +99,9 @@ elif [ ${MODE} = "benchmark_train" ];then ...@@ -99,9 +99,9 @@ elif [ ${MODE} = "benchmark_train" ];then
cd ../../ cd ../../
else else
# prepare lite benchmark coco data # prepare lite benchmark coco data
wget -nc -P ./dataset/coco/ https://paddledet.bj.bcebos.com/data/coco_benchmark.tar --no-check-certificate wget -nc -P ./dataset/coco/ https://bj.bcebos.com/v1/paddledet/data/cocomini.zip --no-check-certificate
cd ./dataset/coco/ && tar -xf coco_benchmark.tar cd ./dataset/coco/ && unzip cocomini.zip
mv -u coco_benchmark/* ./ mv -u cocomini/* ./
ls ./ ls ./
cd ../../ cd ../../
# prepare lite benchmark mot data # prepare lite benchmark mot data
......
...@@ -311,7 +311,7 @@ else ...@@ -311,7 +311,7 @@ else
fi fi
# run train # run train
train_log_path="${LOG_PATH}/${trainer}_gpus_${gpu}_autocast_${autocast}_nodes_${nodes}.log" train_log_path="${LOG_PATH}/${trainer}_gpus_${gpu}_autocast_${autocast}_nodes_${nodes}.log"
eval "${cmd} > ${train_log_path} 2>&1" eval "timeout 5m ${cmd} > ${train_log_path} 2>&1"
last_status=$? last_status=$?
cat ${train_log_path} cat ${train_log_path}
status_check $last_status "${cmd}" "${status_log}" "${model_name}" "${train_log_path}" status_check $last_status "${cmd}" "${status_log}" "${model_name}" "${train_log_path}"
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册