1. 14 10月, 2021 1 次提交
  2. 21 9月, 2021 1 次提交
    • A
      Reuse OneDNN handler for SGD and SUM for SelectedRows input tensors. (#35510) · 799f3861
      Adam Osewski 提交于
      * Create stateful OneDNNAXPYHandler object.
      
      This makes it possible to call it multiple times without recreating the
      oneDNN primitives every time.
      
      * Prepare SGDOpKernel to reuse its implementation from OneDNN kernel.
      
      * OneDNN SGD kernel.
      
      * Update call to use new OneDNNAXPYHandler object api.
      
      * Setup seed in proper place.
      
      * Enable OneDNN kernel only for single case.
      
      * For dense param and sparse grad.
      
      * Small refactor.
      
      * Enable oneDNN by op attr or by cmd line flag.
      
      * Use int64_t type for number of elements.
      
      * Support dense param and grad from OneDNN kernel.
      
      * Enable SGD OneDNN kernel when use MP BF16 optimizer.
      
      * Force non-copyable/movable OneDNNAXPYHandler.
      
      * Reuse OneDNNAXPYHandler for spare tensors in SUM op.
      
      * Fix SFINAE rules.
      
      * Remove recording event inside AXPY.
      
      * Get rid of internal primitive caching.
      
      * Stop use PP cache mechanims to store mem and primitive obj.
      * Handler obj store and reuse needed desc & prim
      
      * Do not derive from MKLDNNHandlerT
      799f3861
  3. 10 9月, 2021 1 次提交
  4. 24 8月, 2021 1 次提交
  5. 17 8月, 2021 1 次提交
  6. 05 8月, 2021 1 次提交
  7. 22 7月, 2021 2 次提交
  8. 19 7月, 2021 1 次提交
  9. 16 7月, 2021 1 次提交
  10. 05 7月, 2021 1 次提交
  11. 01 7月, 2021 1 次提交
  12. 29 6月, 2021 1 次提交
  13. 21 6月, 2021 1 次提交
  14. 16 6月, 2021 1 次提交
  15. 10 6月, 2021 1 次提交
  16. 26 5月, 2021 1 次提交
  17. 07 5月, 2021 1 次提交
  18. 28 4月, 2021 1 次提交
  19. 23 4月, 2021 1 次提交
    • L
      [NPU] refactor check_finite_and_scale npu kernel (#32407) · 39a59dcf
      Leo Chen 提交于
      * refactor_check_finite_and_scale_npu_kernel
      
      * fix compile
      
      * add alloc_float_status op
      
      * add alloc_float_status op
      
      * add FloatStatus for check_finite_and_unscale
      
      * refine code
      
      * remove unneccessary logic
      
      * refine for fleet
      39a59dcf
  20. 22 4月, 2021 1 次提交
  21. 21 4月, 2021 2 次提交
  22. 15 4月, 2021 1 次提交
  23. 08 4月, 2021 1 次提交
  24. 26 3月, 2021 1 次提交
  25. 22 3月, 2021 1 次提交
  26. 20 1月, 2021 1 次提交
  27. 13 1月, 2021 1 次提交
  28. 08 1月, 2021 1 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf
  29. 05 1月, 2021 1 次提交
  30. 15 12月, 2020 1 次提交
  31. 09 12月, 2020 1 次提交
  32. 02 12月, 2020 2 次提交
    • Z
      Add pure fp16 training with master weights. (#27712) · be3777a5
      Zhen Wang 提交于
      * add the weight decay func for the momentum op
      
      * Add the multi_precision function in Momentum Optimizer.
      
      * Make sure that the initial value of master weights are same with the fp16 weights.
      
      * add static loss scaling.
      
      * add the rescale_grad function in the pure fp16 training.
      
      * use the original momentum updating method.
      
      * Polish some codes, such as variable names.
      
      * add docstring for apis.
      
      * update the var creation details of _create_master_weight.
      
      * not modify codes about imperative momentum updating.
      
      * Fix the error of test_dist_sparse_tensor_load_momentum UT.
      
      * add unit test for multi precision fp16 training.
      
      * add more unit tests for CI.
      
      * Use lower threshold values for allclose comparing in test_multi_precision_fp16_train UT.
      
      * For CI Coverage Checking.
      be3777a5
    • F
      Layer norm fp16 (#29169) · 7584bb50
      furnace 提交于
      * add fp16 for layer_norm op
      
      * revert layernorm api
      
      * fix forward
      
      * fix forward
      
      * fix backward for layernorm with fp16
      
      * fix unit test for layernorm with fp16
      
      * fix with_mkldnn compile error for layernorm with fp16
      
      * 1. revert to PADDLE_ENFORCE_NOT_NULL, 2. change static_cast<float> to static_cast<U>
      
      * fix with_mkldnn compile error for layernorm with fp16
      
      * fix with_mkldnn compile error for layernorm with fp16
      Co-authored-by: Nzhiqiu <chenqiuliang@baidu.com>
      7584bb50
  33. 30 11月, 2020 1 次提交
  34. 18 11月, 2020 1 次提交
  35. 04 11月, 2020 1 次提交
  36. 12 10月, 2020 1 次提交
  37. 23 9月, 2020 1 次提交