rec_mv3_tps_bilstm_attn.yml 1.1 KB
Newer Older
L
LDOUBLEV 已提交
1 2
Global:
  algorithm: RARE
T
tink2123 已提交
3
  use_gpu: true
L
LDOUBLEV 已提交
4 5 6
  epoch_num: 72
  log_smooth_window: 20
  print_batch_step: 10
L
LDOUBLEV 已提交
7
  save_model_dir: output/rec_RARE
L
LDOUBLEV 已提交
8 9
  save_epoch_step: 3
  eval_batch_step: 2000
T
tink2123 已提交
10 11
  train_batch_size_per_card: 256
  test_batch_size_per_card: 256
L
LDOUBLEV 已提交
12 13
  image_shape: [3, 32, 100]
  max_text_length: 25
T
tink2123 已提交
14
  character_type: en
L
LDOUBLEV 已提交
15
  loss_type: attention
T
tink2123 已提交
16
  tps: true
L
LDOUBLEV 已提交
17
  reader_yml: ./configs/rec/rec_benchmark_reader.yml
18 19 20
  pretrain_weights:
  checkpoints:
  save_inference_dir:
T
tink2123 已提交
21 22 23
  infer_img:


L
LDOUBLEV 已提交
24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54
Architecture:
  function: ppocr.modeling.architectures.rec_model,RecModel

TPS:
  function: ppocr.modeling.stns.tps,TPS
  num_fiducial: 20
  loc_lr: 0.1
  model_name: small
  
Backbone:
  function: ppocr.modeling.backbones.rec_mobilenet_v3,MobileNetV3
  scale: 0.5
  model_name: large
 
Head:
  function: ppocr.modeling.heads.rec_attention_head,AttentionPredict
  encoder_type: rnn
  SeqRNN:
    hidden_size: 96
  Attention:
    decoder_size: 96
    word_vector_dim: 96
  
Loss:
  function: ppocr.modeling.losses.rec_attention_loss,AttentionLoss
  
Optimizer:
  function: ppocr.optimizer,AdamDecay
  base_lr: 0.001
  beta1: 0.9
  beta2: 0.999