1. 11 7月, 2023 1 次提交
    • P
      support sharding parallel (#54634) · b7a05057
      pangengzheng 提交于
      * support sharding parallel
      
      * fix name
      
      * fix
      
      * update
      
      * test amp for sharding
      
      ---------
      
      Co-authored-by: pangengzheng <pangengzheng.baidu.com>
      b7a05057
  2. 06 7月, 2023 1 次提交
  3. 12 6月, 2023 1 次提交
  4. 09 6月, 2023 1 次提交
  5. 08 6月, 2023 1 次提交
    • Y
      [AMP] Add check_numerics API. (#54301) · a5444592
      Yiqun Liu 提交于
      * Add outputs to check_numerics_kernel.
      
      * Add check_numerics to yaml.
      
      * Add API and unittest.
      
      * Add check_nan_inf_level as argument of check_numerics_kernel.
      
      * Add more unittests.
      
      * Fix static API implementation and unittest.
      
      * Move the implementation of check_numerics to paddle.amp.
      
      * Fix import error.
      a5444592
  6. 30 5月, 2023 1 次提交
    • Y
      [AMP] Reimplement check_nan_inf as check_numerics_kernel. (#52245) · 44bd5927
      Yiqun Liu 提交于
      * Reimplement the check_nan_inf function as check_numerics kernel.
      
      * Remove the cpu implemention to phi.
      
      * Add ifdef for the including of omp.h.
      
      * Move the use of FLAGS_check_nan_inf_level out of header file.
      
      * Implement a common PrintAndThrowError function.
      
      * Fix the error using of __NVCC__, which should be instead with __CUDA_ARCH__.
      
      * Add dependency of phi.
      
      * Polish codes and unittest.
      44bd5927
  7. 22 5月, 2023 2 次提交
    • M
      [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode() (#53856) · 3794d171
      Meteor Liu 提交于
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * fixed cyclic reference that caused patial import
      
      * fixed bad change
      
      * fix bad import
      
      * fix bad import
      
      * fix bad import
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fixed usage of in_dynamic_mode() or in_dygraph_mode()
      
      * revert python3 to python in .pre-commit-config.yaml
      
      * fix merge conflicts
      3794d171
    • Z
      remove depthwise_conv from extra black list (#53901) · 98f4446a
      Zhang Ting 提交于
      98f4446a
  8. 18 5月, 2023 1 次提交
  9. 16 5月, 2023 1 次提交
  10. 11 5月, 2023 1 次提交
  11. 10 5月, 2023 1 次提交
  12. 09 5月, 2023 1 次提交
  13. 08 5月, 2023 1 次提交
  14. 27 4月, 2023 1 次提交
  15. 24 4月, 2023 3 次提交
  16. 18 4月, 2023 1 次提交
  17. 14 4月, 2023 1 次提交
    • Y
      [AMP] Unify the static amp codes of fp16 and bf16. (#52694) · dfcba7f4
      Yiqun Liu 提交于
      * Unify the static amp codes of fp16 and bf16.
      
      * Polish apis and add unittest.
      
      * Add operator stats collecting tools for program.
      
      * Add the check of number of bloat16 operators in unittest.
      
      * Add warning for operator not supported for amp.
      
      * Add testing of BF16 O1 and O2.
      dfcba7f4
  18. 13 4月, 2023 1 次提交
  19. 12 4月, 2023 2 次提交
  20. 10 4月, 2023 1 次提交
  21. 06 4月, 2023 1 次提交
    • K
      rem is_compiled_with_npu (#52385) · 7976e2a3
      Kim Yann 提交于
      * rem is_compiled_with_npu
      
      * rem nup related code
      
      * make lint happy
      
      * rem test
      
      * remove some tests
      
      * Update grad_scaler.py
      
      * fix an error
      7976e2a3
  22. 03 4月, 2023 1 次提交
  23. 30 3月, 2023 3 次提交
  24. 15 3月, 2023 1 次提交
  25. 09 3月, 2023 1 次提交
    • G
      Fix hybrid parallel training strategy using bf16 (#51103) · 8db15a42
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * Remove climits.
      
      * Fix bug of hybrid parallel strategy with recompute using bf16.
      
      * Fix bug of recompute_hybrid ctx.amp_dtype
      
      * Fix bug of amp_dtype.
      
      * Fix bug of auto_cast.
      8db15a42
  26. 24 2月, 2023 1 次提交
    • W
      Revert grad scale optimization pr (#50839) · 8a503522
      Weilong Wu 提交于
      * Revert "fixoptminizer _set_auxiliary_var bug (#50335)"
      
      This reverts commit c44005f0.
      
      * Revert "refine optimizer create accumulators (#50188)"
      
      This reverts commit 244e7546.
      
      * Revert "fix found_inf bug for custom optimizer (#50158)"
      
      This reverts commit 64573f9f.
      
      * Revert "refine amp scaler found_inf (#49864)"
      
      This reverts commit 382e9a06.
      
      * fix code format
      
      * fix conflict
      8a503522
  27. 13 2月, 2023 1 次提交
  28. 03 2月, 2023 1 次提交
  29. 30 1月, 2023 1 次提交
  30. 19 1月, 2023 1 次提交
    • J
      [KUNLUN] add op: maxpool_with_index (#49505) · f71f77e9
      jameszhang 提交于
      * [KUNLUN] add op: maxpool_with_index
      
      * use DeviceContext::Alloc() instead of DenseTensor::mutable_data()
      
      * fix file format
      
      * solve clip unittest failure
      
      * minor fix
      
      * Revert "solve clip unittest failure" since the issue is fixed
      in #49535
      
      This reverts commit 1127adc66e79afe35ac3c00bb34e6aaa7cd7d78b.
      
      * align with xdnn on the definition of mask in max_pool_with_index
      
      * minor
      f71f77e9
  31. 12 1月, 2023 1 次提交
  32. 11 1月, 2023 1 次提交
  33. 06 1月, 2023 1 次提交
  34. 05 1月, 2023 1 次提交