1. 22 5月, 2023 1 次提交
    • M
      [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode() (#53856) · 3794d171
      Meteor Liu 提交于
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * [dygraph]unify _non_static_mode() in_dygraph_mode() and in_dynamic_mode()
      
      * fixed cyclic reference that caused patial import
      
      * fixed bad change
      
      * fix bad import
      
      * fix bad import
      
      * fix bad import
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fix ut failed caused by change in_dynamic_mode
      
      * fixed usage of in_dynamic_mode() or in_dygraph_mode()
      
      * revert python3 to python in .pre-commit-config.yaml
      
      * fix merge conflicts
      3794d171
  2. 18 5月, 2023 1 次提交
    • shaojie_wang's avatar
      [AMP]Master grad in static graph (#53362) · 972581d8
      shaojie_wang 提交于
      * add master gradients on static graph
      
      * add unit test for bf16 master grad static graph
      
      * use float16 as v100 test dtype
      
      * only skip GPU which do not support bf16
      
      * use linear layer to test master grad
      
      * 1.push master grad creation before all optimizer ops; 2.remove useless unittest; 3.use a function to create master grad states
      972581d8
  3. 11 5月, 2023 1 次提交
  4. 27 4月, 2023 1 次提交
  5. 04 4月, 2023 1 次提交
  6. 23 3月, 2023 1 次提交
  7. 20 3月, 2023 1 次提交
  8. 15 3月, 2023 1 次提交
  9. 10 3月, 2023 1 次提交
  10. 28 2月, 2023 1 次提交
  11. 24 2月, 2023 1 次提交
    • W
      Revert grad scale optimization pr (#50839) · 8a503522
      Weilong Wu 提交于
      * Revert "fixoptminizer _set_auxiliary_var bug (#50335)"
      
      This reverts commit c44005f0.
      
      * Revert "refine optimizer create accumulators (#50188)"
      
      This reverts commit 244e7546.
      
      * Revert "fix found_inf bug for custom optimizer (#50158)"
      
      This reverts commit 64573f9f.
      
      * Revert "refine amp scaler found_inf (#49864)"
      
      This reverts commit 382e9a06.
      
      * fix code format
      
      * fix conflict
      8a503522
  12. 06 2月, 2023 1 次提交
  13. 30 1月, 2023 1 次提交
  14. 03 1月, 2023 1 次提交
  15. 30 12月, 2022 1 次提交
  16. 25 12月, 2022 1 次提交
  17. 09 12月, 2022 1 次提交
  18. 29 11月, 2022 1 次提交
  19. 17 11月, 2022 1 次提交
  20. 23 10月, 2022 1 次提交
  21. 23 9月, 2022 1 次提交
  22. 14 9月, 2022 1 次提交
  23. 26 8月, 2022 1 次提交
  24. 15 8月, 2022 1 次提交
  25. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  26. 01 6月, 2022 1 次提交
  27. 15 4月, 2022 1 次提交
  28. 28 3月, 2022 1 次提交
  29. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  30. 27 10月, 2021 1 次提交
  31. 18 10月, 2021 1 次提交
  32. 29 9月, 2021 1 次提交
  33. 26 9月, 2021 1 次提交
  34. 22 9月, 2021 1 次提交
  35. 17 9月, 2021 1 次提交
    • Z
      [AMP] Support pure fp16 training mode for dygraph (#35521) · adaeee4d
      zhangbo9674 提交于
      * add pure fp16 major function in auto_cast & tracer
      
      * support master weight in dygraph for pure fp16
      
      * check mix dtype of fp16&fp32 for check_finite_and_unscale op
      
      * change pure fp16 funtion name
      
      * refine some bug in auto_cast
      
      * refine auto_cast interface logic
      
      * add param _casted_by_pure_fp16 for class Layer
      
      * support state_dict hook for save model by user appointed dtype in pure_fp16_decorator
      
      * refine pure_fp16_decorator as decorator
      
      * add unittest
      
      * add comment
      
      * add comment
      
      * support recompute
      
      * add comment for auto_cast and decorator
      
      * support to_static_state_dict for paddle.jit.save
      
      * unlimite models num and optimizers num
      
      * add lookup_table in black_list
      
      * fix momentum and layer state_dict
      
      * fix bug in layer state_dict
      
      * fix bug in layer state_dict_helper
      
      * refine unittest
      
      * refine test_momentun_op
      
      * refine interface and some code
      
      * refine amp_decorator interface
      
      * refine pure fp16 interface
      
      * refine master weight interface
      adaeee4d
  36. 14 9月, 2021 1 次提交
  37. 01 9月, 2021 1 次提交
  38. 23 8月, 2021 1 次提交
  39. 17 8月, 2021 1 次提交
  40. 02 8月, 2021 1 次提交