1. 15 3月, 2022 1 次提交
  2. 02 12月, 2021 1 次提交
  3. 22 9月, 2021 1 次提交
  4. 17 9月, 2021 1 次提交
    • Z
      [AMP] Support pure fp16 training mode for dygraph (#35521) · adaeee4d
      zhangbo9674 提交于
      * add pure fp16 major function in auto_cast & tracer
      
      * support master weight in dygraph for pure fp16
      
      * check mix dtype of fp16&fp32 for check_finite_and_unscale op
      
      * change pure fp16 funtion name
      
      * refine some bug in auto_cast
      
      * refine auto_cast interface logic
      
      * add param _casted_by_pure_fp16 for class Layer
      
      * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator
      
      * refine pure_fp16_decorator as decorator
      
      * add unittest
      
      * add comment
      
      * add comment
      
      * support recompute
      
      * add comment for auto_cast and decorator
      
      * support to_static_state_dict for paddle.jit.save
      
      * unlimite models num and optimizers num
      
      * add lookup_table in black_list
      
      * fix momentum and layer state_dict
      
      * fix bug in layer state_dict
      
      * fix bug in layer state_dict_helper
      
      * refine unittest
      
      * refine test_momentun_op
      
      * refine interface and some code
      
      * refine amp_decorator interface
      
      * refine pure fp16 interface
      
      * refine master weight interface
      adaeee4d
  5. 16 8月, 2021 1 次提交
  6. 11 8月, 2021 1 次提交
    • Z
      [AMP] add state_dict and load_state_dict and unittest for class GradScaler (#34300) · 99f8f5c8
      zhangbo9674 提交于
      * add state_dict and load_state_dict and unittest for class GradScaler
      
      * refine unittest for coverage of load_state_dict
      
      * refine comments of code-block
      
      * refine some comments
      
      * refine state_dict code and unittest
      
      * add #require gpu, xpu for GradScaler get/set example code
      
      * add #require gpu, xpu for GradScaler get/set example code
      
      * refine example code
      
      * refine unittest for state_dict
      
      * refine unittest for state_dict
      
      * fix bug of DataLoader in TestGradScalerStateDict
      
      * add flag FLAGS_cudnn_deterministic
      99f8f5c8
  7. 15 7月, 2021 1 次提交
  8. 01 7月, 2021 1 次提交
  9. 29 6月, 2021 1 次提交
  10. 14 9月, 2020 1 次提交
    • Z
      Update amp_check_finite_and_scale_op and add an updating_loss_scaling op for... · d708b210
      Zhen Wang 提交于
      Update amp_check_finite_and_scale_op and add an updating_loss_scaling op for static graph amp training. (#26240)
      
      * update amp_check_finite_and_scale_op for static_amp.
      
      * use amp_check_finite_and_scale in static graph amp.
      
      * update grads to zero when grads own infinite values(as for amp_checkout_finite_and_scale op).
      
      * add update_loss_scaling op in cpp.
      
      * add update_loss_scaling_op unit test.
      
      * update the doc of the check_finite_and_unscale op
      
      * Update the process of gradients updating skipping if the gradients have infinite values.
      
      * update the way to zero grads.
      
      * update test_update_loss_scaling_op.py
      
      * add log info when find infinite grads.
      
      * add the unit test for UpdateLossScaling Layer.
      d708b210
  11. 13 8月, 2020 1 次提交
    • L
      Feature/Enable Auto-Mixed-Precision in dynamic graph (#24903) · 2d95280e
      Leo Chen 提交于
      * add auto_cast, test=develop
      
      * add loss scaler, test=develop
      
      * add comments, test=develop
      
      * refine code, test=develop
      
      * refine code, test=develop
      
      * do not set flags automatically, test=develop
      
      * fix custom op bug, test=develop
      
      * add more test, test=develop
      
      * refine enable logic, test=develop
      
      * enable amp test with GPU, test=develop
      
      * add unittest
      
      * add test for found_inf
      
      * follow comments
      
      * follow comments
      
      * remove global variable, use singleton
      
      * add some notes
      
      * update comments
      
      * update comments
      
      * update comments
      
      * add use_dynamic_loss_scaling argument
      
      * refine found_inf
      
      * refine found_inf
      2d95280e