1. 29 4月, 2022 1 次提交
    • W
      [cherry-pick 2.3] Add fused_multi_transformer op to optimize transformer... · 50bfe420
      WangXi 提交于
      [cherry-pick 2.3] Add fused_multi_transformer op to optimize transformer generation performance (#42311)
      
      * Add fused_multi_transformer op to optimize transformer generation performance (#41814)
      
      * fix fused_multi_transformer compile failed in cuda arch < sm53 (#42315)
      
      * fix ci timeout
      50bfe420
  2. 22 4月, 2022 1 次提交
  3. 16 3月, 2022 1 次提交
  4. 19 2月, 2022 1 次提交
    • S
      Add the DistributedFusedLamb optimizer (#39148) · 5df3cd61
      sneaxiy 提交于
      * add DistributedFusedLamb op
      
      * polish code
      
      * fix compile error
      
      * compatible with pten changement
      
      * fix rocm compile error
      
      * improve converage
      
      * update upstream/develop
      
      * fix cast_with_ptr.h
      
      * add FLAGS_distributed_lamb_divide_nranks_when_allreduce=1
      
      * fix clip before allreduce
      
      * add use_master_param_norm
      
      * code polish
      
      * fix bug
      
      * fix ROCM ci
      5df3cd61
  5. 07 2月, 2022 1 次提交
  6. 13 1月, 2022 1 次提交
    • J
      Added mul BF16/FP32 FWD/BWD oneDNN kernel (#38552) · fc6eed5b
      jakpiase 提交于
      * base changes for mul reimplementation
      
      * empty commit
      
      * tmp save
      
      * full implementation of mul bf16/fp32 fwd bwd
      
      * CI fix
      
      * CI rerun
      
      * changed unity build cmake to avoid gpu issues
      
      * removed mul mkldnn from unity build
      
      * added skipping tests if not cpu_bf16
      
      * CI fix
      
      * CI fix
      
      * CI fix
      fc6eed5b
  7. 28 12月, 2021 1 次提交
  8. 20 12月, 2021 1 次提交
    • S
      Support FP16 for more ops (#38123) · 1f445bf3
      sneaxiy 提交于
      * support FP16 for more ops
      
      * add amp list tests
      
      * refine reduce_mean_grad
      
      * fix OP benchmark ci
      
      * fix fp16 reduce_mean
      
      * updat ut, but still have some problems
      
      * remove mean/reduce_mean fp16 kernel
      1f445bf3
  9. 17 12月, 2021 1 次提交
    • S
      Refine some AMP operators for BERT (#37923) · d80fe268
      sneaxiy 提交于
      * support multi precision update for LAMB
      
      * hide some api
      
      * fix ci uts
      
      * fix lamb output of dygraph
      
      * remove some changes to some PR
      
      * try to fix Py3 CI compile error
      
      * fix test_imperative_optimizer, add lars ut, add layer_norm ut
      
      * fix ut, fix format
      
      * fix ut
      
      * fix windows ci
      d80fe268
  10. 27 10月, 2021 1 次提交
  11. 14 10月, 2021 1 次提交
  12. 21 9月, 2021 1 次提交
    • A
      Reuse OneDNN handler for SGD and SUM for SelectedRows input tensors. (#35510) · 799f3861
      Adam Osewski 提交于
      * Create stateful OneDNNAXPYHandler object.
      
      This makes it possible to call it multiple times without recreating the
      oneDNN primitives every time.
      
      * Prepare SGDOpKernel to reuse its implementation from OneDNN kernel.
      
      * OneDNN SGD kernel.
      
      * Update call to use new OneDNNAXPYHandler object api.
      
      * Setup seed in proper place.
      
      * Enable OneDNN kernel only for single case.
      
      * For dense param and sparse grad.
      
      * Small refactor.
      
      * Enable oneDNN by op attr or by cmd line flag.
      
      * Use int64_t type for number of elements.
      
      * Support dense param and grad from OneDNN kernel.
      
      * Enable SGD OneDNN kernel when use MP BF16 optimizer.
      
      * Force non-copyable/movable OneDNNAXPYHandler.
      
      * Reuse OneDNNAXPYHandler for spare tensors in SUM op.
      
      * Fix SFINAE rules.
      
      * Remove recording event inside AXPY.
      
      * Get rid of internal primitive caching.
      
      * Stop use PP cache mechanims to store mem and primitive obj.
      * Handler obj store and reuse needed desc & prim
      
      * Do not derive from MKLDNNHandlerT
      799f3861
  13. 10 9月, 2021 1 次提交
  14. 24 8月, 2021 1 次提交
  15. 17 8月, 2021 1 次提交
  16. 05 8月, 2021 1 次提交
  17. 22 7月, 2021 2 次提交
  18. 19 7月, 2021 1 次提交
  19. 16 7月, 2021 1 次提交
  20. 05 7月, 2021 1 次提交
  21. 01 7月, 2021 1 次提交
  22. 29 6月, 2021 1 次提交
  23. 21 6月, 2021 1 次提交
  24. 16 6月, 2021 1 次提交
  25. 10 6月, 2021 1 次提交
  26. 26 5月, 2021 1 次提交
  27. 07 5月, 2021 1 次提交
  28. 28 4月, 2021 1 次提交
  29. 23 4月, 2021 1 次提交
    • L
      [NPU] refactor check_finite_and_scale npu kernel (#32407) · 39a59dcf
      Leo Chen 提交于
      * refactor_check_finite_and_scale_npu_kernel
      
      * fix compile
      
      * add alloc_float_status op
      
      * add alloc_float_status op
      
      * add FloatStatus for check_finite_and_unscale
      
      * refine code
      
      * remove unneccessary logic
      
      * refine for fleet
      39a59dcf
  30. 22 4月, 2021 1 次提交
  31. 21 4月, 2021 2 次提交
  32. 15 4月, 2021 1 次提交
  33. 08 4月, 2021 1 次提交
  34. 26 3月, 2021 1 次提交
  35. 22 3月, 2021 1 次提交
  36. 20 1月, 2021 1 次提交
  37. 13 1月, 2021 1 次提交
  38. 08 1月, 2021 1 次提交
    • Z
      Support pure fp16 training for AMP API. (#29544) · 7f7dfccf
      Zhen Wang 提交于
      * add cast ops before and after unsupported fp16 ops.
      
      * Keep partial net in FP32 pattern.
      
      * Support check_finite_and_unscale and update_loss_scaling for FP16 calculation mode.
      
      * Add fp16 support for adam op.
      
      * add multi precision attr for adam.
      
      * Fix the bug of test_multi_precision_fp16_train UT.
      
      * Code format for CI.
      
      * Fix the redefine error about MPTypeTrait on windows.
      
      * fix bugs of the _create_accumulators func in Momentum.
      
      * fix bug when inserting post cast op.
      
      * Add the update_loss_scaling op in allow_set of UnusedVarCheck.
      
      * Update for ci coverage.
      
      * Add some doc for OptimizerWithMixedPrecision.
      
      * Fix the code style.
      
      * Imporve the doc of `amp_init`.
      
      * Change for fp16 testing if users have the infer program defined in separate way.
      7f7dfccf