1. 12 6月, 2023 1 次提交
  2. 16 5月, 2023 1 次提交
  3. 11 5月, 2023 1 次提交
  4. 10 5月, 2023 1 次提交
  5. 08 5月, 2023 1 次提交
  6. 24 4月, 2023 1 次提交
  7. 14 4月, 2023 1 次提交
    • Y
      [AMP] Unify the static amp codes of fp16 and bf16. (#52694) · dfcba7f4
      Yiqun Liu 提交于
      * Unify the static amp codes of fp16 and bf16.
      
      * Polish apis and add unittest.
      
      * Add operator stats collecting tools for program.
      
      * Add the check of number of bloat16 operators in unittest.
      
      * Add warning for operator not supported for amp.
      
      * Add testing of BF16 O1 and O2.
      dfcba7f4
  8. 06 4月, 2023 1 次提交
    • K
      rem is_compiled_with_npu (#52385) · 7976e2a3
      Kim Yann 提交于
      * rem is_compiled_with_npu
      
      * rem nup related code
      
      * make lint happy
      
      * rem test
      
      * remove some tests
      
      * Update grad_scaler.py
      
      * fix an error
      7976e2a3
  9. 03 4月, 2023 1 次提交
  10. 17 1月, 2023 1 次提交
  11. 12 1月, 2023 1 次提交
  12. 08 11月, 2022 1 次提交
  13. 23 10月, 2022 1 次提交
  14. 29 8月, 2022 1 次提交
  15. 26 4月, 2022 1 次提交
  16. 16 3月, 2022 1 次提交
  17. 28 12月, 2021 1 次提交
  18. 20 12月, 2021 1 次提交
    • S
      Support FP16 for more ops (#38123) · 1f445bf3
      sneaxiy 提交于
      * support FP16 for more ops
      
      * add amp list tests
      
      * refine reduce_mean_grad
      
      * fix OP benchmark ci
      
      * fix fp16 reduce_mean
      
      * updat ut, but still have some problems
      
      * remove mean/reduce_mean fp16 kernel
      1f445bf3
  19. 27 10月, 2021 1 次提交
  20. 10 9月, 2021 1 次提交
  21. 05 8月, 2021 1 次提交
  22. 22 7月, 2021 1 次提交
  23. 05 7月, 2021 1 次提交
  24. 01 7月, 2021 1 次提交
  25. 29 6月, 2021 1 次提交
  26. 21 6月, 2021 1 次提交
  27. 26 5月, 2021 1 次提交
  28. 08 4月, 2021 1 次提交
  29. 22 3月, 2021 1 次提交
  30. 20 1月, 2021 1 次提交
  31. 13 1月, 2021 1 次提交
  32. 08 1月, 2021 1 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf
  33. 02 12月, 2020 1 次提交
    • F
      Layer norm fp16 (#29169) · 7584bb50
      furnace 提交于
      * add fp16 for layer_norm op
      
      * revert layernorm api
      
      * fix forward
      
      * fix forward
      
      * fix backward for layernorm with fp16
      
      * fix unit test for layernorm with fp16
      
      * fix with_mkldnn compile error for layernorm with fp16
      
      * 1. revert to PADDLE_ENFORCE_NOT_NULL, 2. change static_cast<float> to static_cast<U>
      
      * fix with_mkldnn compile error for layernorm with fp16
      
      * fix with_mkldnn compile error for layernorm with fp16
      Co-authored-by: Nzhiqiu <chenqiuliang@baidu.com>
      7584bb50
  34. 18 11月, 2020 1 次提交
  35. 23 9月, 2020 1 次提交
  36. 26 11月, 2019 1 次提交
  37. 30 10月, 2019 1 次提交
  38. 19 9月, 2019 1 次提交
  39. 06 9月, 2019 1 次提交
  40. 31 8月, 2019 1 次提交