1. 06 7月, 2023 1 次提交
  2. 12 6月, 2023 1 次提交
  3. 09 6月, 2023 1 次提交
  4. 08 6月, 2023 1 次提交
    • Y
      [AMP] Add check_numerics API. (#54301) · a5444592
      Yiqun Liu 提交于
      * Add outputs to check_numerics_kernel.
      
      * Add check_numerics to yaml.
      
      * Add API and unittest.
      
      * Add check_nan_inf_level as argument of check_numerics_kernel.
      
      * Add more unittests.
      
      * Fix static API implementation and unittest.
      
      * Move the implementation of check_numerics to paddle.amp.
      
      * Fix import error.
      a5444592
  5. 30 5月, 2023 1 次提交
    • Y
      [AMP] Reimplement check_nan_inf as check_numerics_kernel. (#52245) · 44bd5927
      Yiqun Liu 提交于
      * Reimplement the check_nan_inf function as check_numerics kernel.
      
      * Remove the cpu implemention to phi.
      
      * Add ifdef for the including of omp.h.
      
      * Move the use of FLAGS_check_nan_inf_level out of header file.
      
      * Implement a common PrintAndThrowError function.
      
      * Fix the error using of __NVCC__, which should be instead with __CUDA_ARCH__.
      
      * Add dependency of phi.
      
      * Polish codes and unittest.
      44bd5927
  6. 22 5月, 2023 2 次提交
    • M
      [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode() (#53856) · 3794d171
      Meteor Liu 提交于
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * fixed cyclic reference that caused patial import
      
      * fixed bad change
      
      * fix bad import
      
      * fix bad import
      
      * fix bad import
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fixed usage of in_dynamic_mode() or in_dygraph_mode()
      
      * revert python3 to python in .pre-commit-config.yaml
      
      * fix merge conflicts
      3794d171
    • Z
      remove depthwise_conv from extra black list (#53901) · 98f4446a
      Zhang Ting 提交于
      98f4446a
  7. 18 5月, 2023 1 次提交
  8. 16 5月, 2023 1 次提交
  9. 11 5月, 2023 1 次提交
  10. 10 5月, 2023 1 次提交
  11. 09 5月, 2023 1 次提交
  12. 08 5月, 2023 1 次提交
  13. 27 4月, 2023 1 次提交
  14. 24 4月, 2023 3 次提交
  15. 18 4月, 2023 1 次提交
  16. 14 4月, 2023 1 次提交
    • Y
      [AMP] Unify the static amp codes of fp16 and bf16. (#52694) · dfcba7f4
      Yiqun Liu 提交于
      * Unify the static amp codes of fp16 and bf16.
      
      * Polish apis and add unittest.
      
      * Add operator stats collecting tools for program.
      
      * Add the check of number of bloat16 operators in unittest.
      
      * Add warning for operator not supported for amp.
      
      * Add testing of BF16 O1 and O2.
      dfcba7f4
  17. 13 4月, 2023 1 次提交
  18. 12 4月, 2023 2 次提交
  19. 10 4月, 2023 1 次提交
  20. 06 4月, 2023 1 次提交
    • K
      rem is_compiled_with_npu (#52385) · 7976e2a3
      Kim Yann 提交于
      * rem is_compiled_with_npu
      
      * rem nup related code
      
      * make lint happy
      
      * rem test
      
      * remove some tests
      
      * Update grad_scaler.py
      
      * fix an error
      7976e2a3
  21. 03 4月, 2023 1 次提交
  22. 30 3月, 2023 3 次提交
  23. 15 3月, 2023 1 次提交
  24. 09 3月, 2023 1 次提交
    • G
      Fix hybrid parallel training strategy using bf16 (#51103) · 8db15a42
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * Remove climits.
      
      * Fix bug of hybrid parallel strategy with recompute using bf16.
      
      * Fix bug of recompute_hybrid ctx.amp_dtype
      
      * Fix bug of amp_dtype.
      
      * Fix bug of auto_cast.
      8db15a42
  25. 24 2月, 2023 1 次提交
    • W
      Revert grad scale optimization pr (#50839) · 8a503522
      Weilong Wu 提交于
      * Revert "fixoptminizer _set_auxiliary_var bug (#50335)"
      
      This reverts commit c44005f0.
      
      * Revert "refine optimizer create accumulators (#50188)"
      
      This reverts commit 244e7546.
      
      * Revert "fix found_inf bug for custom optimizer (#50158)"
      
      This reverts commit 64573f9f.
      
      * Revert "refine amp scaler found_inf (#49864)"
      
      This reverts commit 382e9a06.
      
      * fix code format
      
      * fix conflict
      8a503522
  26. 13 2月, 2023 1 次提交
  27. 03 2月, 2023 1 次提交
  28. 30 1月, 2023 1 次提交
  29. 19 1月, 2023 1 次提交
    • J
      [KUNLUN] add op: maxpool_with_index (#49505) · f71f77e9
      jameszhang 提交于
      * [KUNLUN] add op: maxpool_with_index
      
      * use DeviceContext::Alloc() instead of DenseTensor::mutable_data()
      
      * fix file format
      
      * solve clip unittest failure
      
      * minor fix
      
      * Revert "solve clip unittest failure" since the issue is fixed
      in #49535
      
      This reverts commit 1127adc66e79afe35ac3c00bb34e6aaa7cd7d78b.
      
      * align with xdnn on the definition of mask in max_pool_with_index
      
      * minor
      f71f77e9
  30. 12 1月, 2023 1 次提交
  31. 11 1月, 2023 1 次提交
  32. 06 1月, 2023 1 次提交
  33. 05 1月, 2023 1 次提交
  34. 15 12月, 2022 1 次提交
    • M
      修复paddle.amp.decorate等API的文档 (#48983) · c5af51ca
      mjxs 提交于
      * 涉及到的api有
      paddle.amp.decorate
      paddle.static.npu_places
      paddle.signal.istft
      paddle.signal.stft
      paddle.linalg.eigvalsh
      paddle.randint_like
      
      * change signal.stft
      
      * randint_like的low增加optional
      
      * ; test=docs_preview
      
      * 修改了注解格式; test=docs_preview
      
      * 修改了公式格式
      
      * 修改了decorate的models等
      
      * test=document_fix
      Co-authored-by: NLigoml <39876205+Ligoml@users.noreply.github.com>
      c5af51ca