提交 0f86c555 编写于 作者: G gaotingquan 提交者: Tingquan Gao

add amp args, use_amp=False

上级 2d8346cd
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: AlexNet name: AlexNet
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSPDarkNet53 name: CSPDarkNet53
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_base_224 name: CSWinTransformer_base_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_base_384 name: CSWinTransformer_base_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_large_224 name: CSWinTransformer_large_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_large_384 name: CSWinTransformer_large_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_small_224 name: CSWinTransformer_small_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_tiny_224 name: CSWinTransformer_tiny_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_base_224 name: ConvNeXt_base_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_base_384 name: ConvNeXt_base_384
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_large_224 name: ConvNeXt_large_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_large_384 name: ConvNeXt_large_384
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_small name: ConvNeXt_small
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_tiny name: ConvNeXt_tiny
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_13_224 name: CvT_13_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_13_384 name: CvT_13_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_21_224 name: CvT_21_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_21_384 name: CvT_21_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_W24_384 name: CvT_W24_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102 name: DLA102
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102x name: DLA102x
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102x2 name: DLA102x2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA169 name: DLA169
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA34 name: DLA34
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA46_c name: DLA46_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA46x_c name: DLA46x_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60 name: DLA60
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60x name: DLA60x
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60x_c name: DLA60x_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN107 name: DPN107
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN131 name: DPN131
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN68 name: DPN68
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN92 name: DPN92
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN98 name: DPN98
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_base name: DSNet_base
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_small name: DSNet_small
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_tiny name: DSNet_tiny
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DarkNet53 name: DarkNet53
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_distilled_patch16_224 name: DeiT_base_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_distilled_patch16_384 name: DeiT_base_distilled_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_patch16_224 name: DeiT_base_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_patch16_384 name: DeiT_base_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_small_distilled_patch16_224 name: DeiT_small_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_small_patch16_224 name: DeiT_small_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_tiny_distilled_patch16_224 name: DeiT_tiny_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_tiny_patch16_224 name: DeiT_tiny_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet121 name: DenseNet121
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet161 name: DenseNet161
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet169 name: DenseNet169
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet201 name: DenseNet201
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet264 name: DenseNet264
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: "./inference" save_inference_dir: "./inference"
use_dali: false use_dali: false
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_25 name: ESNet_x0_25
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_5 name: ESNet_x0_5
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_75 name: ESNet_x0_75
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x1_0 name: ESNet_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB0 name: EfficientNetB0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 240, 240] image_shape: [3, 240, 240]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB1 name: EfficientNetB1
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 260, 260] image_shape: [3, 260, 260]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB2 name: EfficientNetB2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 300, 300] image_shape: [3, 300, 300]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB3 name: EfficientNetB3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 380, 380] image_shape: [3, 380, 380]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB4 name: EfficientNetB4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 456, 456] image_shape: [3, 456, 456]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB5 name: EfficientNetB5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 528, 528] image_shape: [3, 528, 528]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB6 name: EfficientNetB6
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 600, 600] image_shape: [3, 600, 600]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB7 name: EfficientNetB7
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x0_5 name: GhostNet_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x1_0 name: GhostNet_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x1_3 name: GhostNet_x1_3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W18_C name: HRNet_W18_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W30_C name: HRNet_W30_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W32_C name: HRNet_W32_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W40_C name: HRNet_W40_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W44_C name: HRNet_W44_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W48_C name: HRNet_W48_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W64_C name: HRNet_W64_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet39_ds name: HarDNet39_ds
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet68 name: HarDNet68
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet68_ds name: HarDNet68_ds
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet85 name: HarDNet85
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GoogLeNet name: GoogLeNet
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: InceptionV3 name: InceptionV3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: InceptionV4 name: InceptionV4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_128 name: LeViT_128
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_128S name: LeViT_128S
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_192 name: LeViT_192
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_256 name: LeViT_256
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_384 name: LeViT_384
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MicroNet_M0 name: MicroNet_M0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MicroNet_M1 name: MicroNet_M1
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MicroNet_M2 name: MicroNet_M2
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MicroNet_M3 name: MicroNet_M3
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MixNet_L name: MixNet_L
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MixNet_M name: MixNet_M
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MixNet_S name: MixNet_S
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNeXt_x1_0 name: MobileNeXt_x1_0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV1 name: MobileNetV1
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV1_x0_25 name: MobileNetV1_x0_25
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV1_x0_5 name: MobileNetV1_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV1_x0_75 name: MobileNetV1_x0_75
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2 name: MobileNetV2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2_x0_25 name: MobileNetV2_x0_25
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2_x0_5 name: MobileNetV2_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2_x0_75 name: MobileNetV2_x0_75
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2_x1_5 name: MobileNetV2_x1_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV2_x2_0 name: MobileNetV2_x2_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_large_x0_35 name: MobileNetV3_large_x0_35
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_large_x0_5 name: MobileNetV3_large_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_large_x0_75 name: MobileNetV3_large_x0_75
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_large_x1_0 name: MobileNetV3_large_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_large_x1_25 name: MobileNetV3_large_x1_25
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x0_35 name: MobileNetV3_small_x0_35
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x0_5 name: MobileNetV3_small_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x0_75 name: MobileNetV3_small_x0_75
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x1_0 name: MobileNetV3_small_x1_0
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x1_0 name: MobileNetV3_small_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileNetV3_small_x1_25 name: MobileNetV3_small_x1_25
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
use_dali: False use_dali: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileViT_S name: MobileViT_S
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
use_dali: False use_dali: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileViT_XS name: MobileViT_XS
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
use_dali: False use_dali: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: MobileViT_XXS name: MobileViT_XXS
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x0_25 name: PPLCNet_x0_25
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x0_35 name: PPLCNet_x0_35
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x0_5 name: PPLCNet_x0_5
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x0_75 name: PPLCNet_x0_75
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x1_0 name: PPLCNet_x1_0
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x1_0 name: PPLCNet_x1_0
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x1_5 name: PPLCNet_x1_5
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x2_0 name: PPLCNet_x2_0
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNet_x2_5 name: PPLCNet_x2_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNetV2_base name: PPLCNetV2_base
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNetV2_large name: PPLCNetV2_large
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PPLCNetV2_small name: PPLCNetV2_small
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B0 name: PVT_V2_B0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B1 name: PVT_V2_B1
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B2 name: PVT_V2_B2
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B2_Linear name: PVT_V2_B2_Linear
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B3 name: PVT_V2_B3
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B4 name: PVT_V2_B4
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PVT_V2_B5 name: PVT_V2_B5
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: PeleeNet name: PeleeNet
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ReXNet_1_0 name: ReXNet_1_0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ReXNet_1_3 name: ReXNet_1_3
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ReXNet_1_5 name: ReXNet_1_5
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ReXNet_2_0 name: ReXNet_2_0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ReXNet_3_0 name: ReXNet_3_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RedNet101 name: RedNet101
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RedNet152 name: RedNet152
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RedNet26 name: RedNet26
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RedNet38 name: RedNet38
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RedNet50 name: RedNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_12GF name: RegNetX_12GF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_1600MF name: RegNetX_1600MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_16GF name: RegNetX_16GF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_200MF name: RegNetX_200MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_3200MF name: RegNetX_3200MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_32GF name: RegNetX_32GF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_400MF name: RegNetX_400MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_600MF name: RegNetX_600MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_6400MF name: RegNetX_6400MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_800MF name: RegNetX_800MF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RegNetX_8GF name: RegNetX_8GF
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_A0 name: RepVGG_A0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_A1 name: RepVGG_A1
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_A2 name: RepVGG_A2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B0 name: RepVGG_B0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B1 name: RepVGG_B1
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B1g2 name: RepVGG_B1g2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B1g4 name: RepVGG_B1g4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B2 name: RepVGG_B2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B2g4 name: RepVGG_B2g4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B3 name: RepVGG_B3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_B3g4 name: RepVGG_B3g4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 320, 320] image_shape: [3, 320, 320]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: RepVGG_D2se name: RepVGG_D2se
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Res2Net101_vd_26w_4s name: Res2Net101_vd_26w_4s
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Res2Net200_vd_26w_4s name: Res2Net200_vd_26w_4s
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Res2Net50_14w_8s name: Res2Net50_14w_8s
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Res2Net50_26w_4s name: Res2Net50_26w_4s
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Res2Net50_vd_26w_4s name: Res2Net50_vd_26w_4s
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeSt101 name: ResNeSt101
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 320, 320] image_shape: [3, 320, 320]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeSt200 name: ResNeSt200
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 416, 416] image_shape: [3, 416, 416]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeSt269 name: ResNeSt269
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeSt50 name: ResNeSt50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeSt50_fast_1s1x64d name: ResNeSt50_fast_1s1x64d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_32x4d name: ResNeXt101_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_64x4d name: ResNeXt101_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_vd_32x4d name: ResNeXt101_vd_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_vd_64x4d name: ResNeXt101_vd_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt152_32x4d name: ResNeXt152_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt152_64x4d name: ResNeXt152_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt152_vd_32x4d name: ResNeXt152_vd_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt152_vd_64x4d name: ResNeXt152_vd_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt50_32x4d name: ResNeXt50_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt50_64x4d name: ResNeXt50_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt50_vd_32x4d name: ResNeXt50_vd_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt50_vd_64x4d name: ResNeXt50_vd_64x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_32x16d_wsl name: ResNeXt101_32x16d_wsl
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_32x32d_wsl name: ResNeXt101_32x32d_wsl
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_32x48d_wsl name: ResNeXt101_32x48d_wsl
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNeXt101_32x8d_wsl name: ResNeXt101_32x8d_wsl
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet101 name: ResNet101
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet101_vd name: ResNet101_vd
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet152 name: ResNet152
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet152_vd name: ResNet152_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet18 name: ResNet18
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet18 name: ResNet18
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet18_vd name: ResNet18_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet200_vd name: ResNet200_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet34 name: ResNet34
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet34_vd name: ResNet34_vd
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50_vd name: ResNet50_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SENet154_vd name: SENet154_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNeXt101_32x4d name: SE_ResNeXt101_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNeXt50_32x4d name: SE_ResNeXt50_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNeXt50_vd_32x4d name: SE_ResNeXt50_vd_32x4d
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNet18_vd name: SE_ResNet18_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNet34_vd name: SE_ResNet34_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SE_ResNet50_vd name: SE_ResNet50_vd
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_swish name: ShuffleNetV2_swish
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x0_25 name: ShuffleNetV2_x0_25
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x0_33 name: ShuffleNetV2_x0_33
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x0_5 name: ShuffleNetV2_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x1_0 name: ShuffleNetV2_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x1_5 name: ShuffleNetV2_x1_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ShuffleNetV2_x2_0 name: ShuffleNetV2_x2_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SqueezeNet1_0 name: SqueezeNet1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SqueezeNet1_1 name: SqueezeNet1_1
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_base_patch4_window12_384 name: SwinTransformer_base_patch4_window12_384
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_base_patch4_window7_224 name: SwinTransformer_base_patch4_window7_224
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_large_patch4_window12_384 name: SwinTransformer_large_patch4_window12_384
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_large_patch4_window7_224 name: SwinTransformer_large_patch4_window7_224
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_small_patch4_window7_224 name: SwinTransformer_small_patch4_window7_224
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformer_tiny_patch4_window7_224 name: SwinTransformer_tiny_patch4_window7_224
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_base_patch4_window16_256 name: SwinTransformerV2_base_patch4_window16_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_base_patch4_window24_384 name: SwinTransformerV2_base_patch4_window24_384
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_base_patch4_window8_256 name: SwinTransformerV2_base_patch4_window8_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_large_patch4_window16_256 name: SwinTransformerV2_large_patch4_window16_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_large_patch4_window24_384 name: SwinTransformerV2_large_patch4_window24_384
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_small_patch4_window16_256 name: SwinTransformerV2_small_patch4_window16_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_small_patch4_window8_256 name: SwinTransformerV2_small_patch4_window8_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_tiny_patch4_window16_256 name: SwinTransformerV2_tiny_patch4_window16_256
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: SwinTransformerV2_tiny_patch4_window8_256 name: SwinTransformerV2_tiny_patch4_window8_256
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TNT_base name: TNT_base
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TNT_small name: TNT_small
......
...@@ -18,6 +18,18 @@ Global: ...@@ -18,6 +18,18 @@ Global:
EMA: EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TinyNet_A name: TinyNet_A
......
...@@ -18,6 +18,18 @@ Global: ...@@ -18,6 +18,18 @@ Global:
EMA: EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TinyNet_B name: TinyNet_B
......
...@@ -18,6 +18,18 @@ Global: ...@@ -18,6 +18,18 @@ Global:
EMA: EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TinyNet_C name: TinyNet_C
......
...@@ -18,6 +18,18 @@ Global: ...@@ -18,6 +18,18 @@ Global:
EMA: EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TinyNet_D name: TinyNet_D
......
...@@ -18,6 +18,18 @@ Global: ...@@ -18,6 +18,18 @@ Global:
EMA: EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: TinyNet_E name: TinyNet_E
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: alt_gvt_base name: alt_gvt_base
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: alt_gvt_large name: alt_gvt_large
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: alt_gvt_small name: alt_gvt_small
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: pcpvt_base name: pcpvt_base
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: pcpvt_large name: pcpvt_large
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: pcpvt_small name: pcpvt_small
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: UniFormer_base name: UniFormer_base
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: UniFormer_base_ls name: UniFormer_base_ls
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: UniFormer_small name: UniFormer_small
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: UniFormer_small_plus name: UniFormer_small_plus
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: UniFormer_small_plus_dim64 name: UniFormer_small_plus_dim64
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VAN_B0 name: VAN_B0
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VAN_B1 name: VAN_B1
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VAN_B2 name: VAN_B2
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VAN_B3 name: VAN_B3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VGG11 name: VGG11
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VGG13 name: VGG13
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VGG16 name: VGG16
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: VGG19 name: VGG19
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_base_patch16_224 name: ViT_base_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_base_patch16_384 name: ViT_base_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_base_patch32_384 name: ViT_base_patch32_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_large_patch16_224 name: ViT_large_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_large_patch16_384 name: ViT_large_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_large_patch32_384 name: ViT_large_patch32_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ViT_small_patch16_224 name: ViT_small_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Xception41 name: Xception41
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Xception41_deeplab name: Xception41_deeplab
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Xception65 name: Xception65
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Xception65_deeplab name: Xception65_deeplab
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: Xception71 name: Xception71
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册