提交 0f86c555 编写于 作者: G gaotingquan 提交者: Tingquan Gao

add amp args, use_amp=False

上级 2d8346cd
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: AlexNet name: AlexNet
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSPDarkNet53 name: CSPDarkNet53
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_base_224 name: CSWinTransformer_base_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_base_384 name: CSWinTransformer_base_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_large_224 name: CSWinTransformer_large_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_large_384 name: CSWinTransformer_large_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_small_224 name: CSWinTransformer_small_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CSWinTransformer_tiny_224 name: CSWinTransformer_tiny_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_base_224 name: ConvNeXt_base_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_base_384 name: ConvNeXt_base_384
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_large_224 name: ConvNeXt_large_224
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_large_384 name: ConvNeXt_large_384
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_small name: ConvNeXt_small
......
...@@ -22,6 +22,18 @@ EMA: ...@@ -22,6 +22,18 @@ EMA:
decay: 0.9999 decay: 0.9999
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ConvNeXt_tiny name: ConvNeXt_tiny
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_13_224 name: CvT_13_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_13_384 name: CvT_13_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_21_224 name: CvT_21_224
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_21_384 name: CvT_21_384
......
...@@ -17,6 +17,18 @@ Global: ...@@ -17,6 +17,18 @@ Global:
to_static: False to_static: False
update_freq: 2 # for 8 cards update_freq: 2 # for 8 cards
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: CvT_W24_384 name: CvT_W24_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102 name: DLA102
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102x name: DLA102x
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA102x2 name: DLA102x2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA169 name: DLA169
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA34 name: DLA34
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA46_c name: DLA46_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA46x_c name: DLA46x_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60 name: DLA60
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60x name: DLA60x
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DLA60x_c name: DLA60x_c
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN107 name: DPN107
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN131 name: DPN131
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN68 name: DPN68
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN92 name: DPN92
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DPN98 name: DPN98
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_base name: DSNet_base
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_small name: DSNet_small
......
...@@ -16,6 +16,18 @@ Global: ...@@ -16,6 +16,18 @@ Global:
# training model under @to_static # training model under @to_static
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DSNet_tiny name: DSNet_tiny
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 256, 256] image_shape: [3, 256, 256]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DarkNet53 name: DarkNet53
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ResNet50 name: ResNet50
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_distilled_patch16_224 name: DeiT_base_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_distilled_patch16_384 name: DeiT_base_distilled_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_patch16_224 name: DeiT_base_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 384, 384] image_shape: [3, 384, 384]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_base_patch16_384 name: DeiT_base_patch16_384
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_small_distilled_patch16_224 name: DeiT_small_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_small_patch16_224 name: DeiT_small_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_tiny_distilled_patch16_224 name: DeiT_tiny_distilled_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DeiT_tiny_patch16_224 name: DeiT_tiny_patch16_224
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet121 name: DenseNet121
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet161 name: DenseNet161
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet169 name: DenseNet169
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet201 name: DenseNet201
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: DenseNet264 name: DenseNet264
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: "./inference" save_inference_dir: "./inference"
use_dali: false use_dali: false
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -15,6 +15,18 @@ Global: ...@@ -15,6 +15,18 @@ Global:
save_inference_dir: ./inference save_inference_dir: ./inference
to_static: False to_static: False
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: "./inference" save_inference_dir: "./inference"
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: "DistillationModel" name: "DistillationModel"
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_25 name: ESNet_x0_25
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_5 name: ESNet_x0_5
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x0_75 name: ESNet_x0_75
......
...@@ -13,6 +13,18 @@ Global: ...@@ -13,6 +13,18 @@ Global:
# used for static mode and model export # used for static mode and model export
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: ESNet_x1_0 name: ESNet_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB0 name: EfficientNetB0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 240, 240] image_shape: [3, 240, 240]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB1 name: EfficientNetB1
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 260, 260] image_shape: [3, 260, 260]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB2 name: EfficientNetB2
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 300, 300] image_shape: [3, 300, 300]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB3 name: EfficientNetB3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 380, 380] image_shape: [3, 380, 380]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB4 name: EfficientNetB4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 456, 456] image_shape: [3, 456, 456]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB5 name: EfficientNetB5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 528, 528] image_shape: [3, 528, 528]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB6 name: EfficientNetB6
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 600, 600] image_shape: [3, 600, 600]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: EfficientNetB7 name: EfficientNetB7
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x0_5 name: GhostNet_x0_5
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x1_0 name: GhostNet_x1_0
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GhostNet_x1_3 name: GhostNet_x1_3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W18_C name: HRNet_W18_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W30_C name: HRNet_W30_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W32_C name: HRNet_W32_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W40_C name: HRNet_W40_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W44_C name: HRNet_W44_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W48_C name: HRNet_W48_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HRNet_W64_C name: HRNet_W64_C
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet39_ds name: HarDNet39_ds
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet68 name: HarDNet68
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet68_ds name: HarDNet68_ds
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: HarDNet85 name: HarDNet85
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: GoogLeNet name: GoogLeNet
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: InceptionV3 name: InceptionV3
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 299, 299] image_shape: [3, 299, 299]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: InceptionV4 name: InceptionV4
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_128 name: LeViT_128
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_128S name: LeViT_128S
......
...@@ -14,6 +14,18 @@ Global: ...@@ -14,6 +14,18 @@ Global:
image_shape: [3, 224, 224] image_shape: [3, 224, 224]
save_inference_dir: ./inference save_inference_dir: ./inference
# mixed precision
AMP:
use_amp: False
use_fp16_test: False
scale_loss: 128.0
use_dynamic_loss_scaling: True
use_promote: False
# O1: mixed fp16, O2: pure fp16
level: O1
# model architecture # model architecture
Arch: Arch:
name: LeViT_192 name: LeViT_192
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册