1. 17 3月, 2023 1 次提交
  2. 16 3月, 2023 1 次提交
    • J
      [Auto Parallel Performance] Support BF16 Training (#51285) · 9ded5707
      JZ-LIANG 提交于
      * update env setting
      
      * update pass logic
      
      * dist op support bf16
      
      * backward cast update
      
      * update setting
      
      * update backward
      
      * revert amp pass
      
      * update fp16 backward logic
      
      * register c_embedding bf16
      
      * revert engine
      
      * add unitest
      
      * add unitest
      
      * update unitest
      
      * update cmake
      
      * update math
      
      * update math.py
      
      * update unitest
      
      * update unitest
      
      * revise unitest
      
      * revise unitest
      
      * update unitest
      
      * update unitest
      
      * update unitest
      9ded5707
  3. 15 3月, 2023 2 次提交
  4. 13 3月, 2023 2 次提交
    • A
      dc956767
    • K
      [with_data_parallel][part6] remove with_data_parallel in distributed optimizer (#50719) · 1404f732
      kangguangli 提交于
      * find relevant testcase
      
      * remove with_data_parallel
      
      * trigger CI
      
      * do not apply ParameterServerGraphOptimizer
      
      * remove useless optimizer
      
      * remove with_data_parallel in test_dist_base
      
      * fix test_fleet_base_3
      
      * only reserve changes for GraphExecutionOptimizer
      
      * fix bug
      
      * fix test_minst_dgc_nccl
      
      * fix typo
      
      * fix test_dist_mnist_gradient_merge
      
      * rm TestDistMnistNCCL2DGCMultiCards
      
      * fix optimizer conflicts
      
      * fix dist_mnist
      
      * fix test_dist_hapi
      
      * delete test_fleet_graph_execution_meta_optimizer & test_fleet_graph_executor
      
      * temporally not delete unittest
      
      * fix unittests
      
      * fix ci
      
      * recover prune in python/paddle/hapi/model.py
      1404f732
  5. 10 3月, 2023 1 次提交
  6. 09 3月, 2023 2 次提交
    • Z
      Remove paddle.fluid.layers.utils.* (#51033) · 86e990d4
      zqw_1997 提交于
      * move fluid.utils to paddle.utils.layers_utils
      
      * fix error
      
      * delete original fluid layers utils
      
      * remove import and old utils
      
      * remove more old utils import
      
      * change import path of fill_constant in the layers_utils.py
      
      * fix mistake
      
      * fix error
      
      * expose in __init__.py
      
      * for comment
      
      * when change the ref of func is_sequence, it should change to the root of is_sequence instead
      
      * for codecheck
      86e990d4
    • G
      Fix hybrid parallel training strategy using bf16 (#51103) · 8db15a42
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * Remove climits.
      
      * Fix bug of hybrid parallel strategy with recompute using bf16.
      
      * Fix bug of recompute_hybrid ctx.amp_dtype
      
      * Fix bug of amp_dtype.
      
      * Fix bug of auto_cast.
      8db15a42
  7. 08 3月, 2023 2 次提交
  8. 06 3月, 2023 3 次提交
  9. 03 3月, 2023 3 次提交
  10. 02 3月, 2023 2 次提交
  11. 01 3月, 2023 1 次提交
  12. 28 2月, 2023 2 次提交
  13. 27 2月, 2023 4 次提交
  14. 24 2月, 2023 1 次提交
    • W
      Revert grad scale optimization pr (#50839) · 8a503522
      Weilong Wu 提交于
      * Revert "fixoptminizer _set_auxiliary_var bug (#50335)"
      
      This reverts commit c44005f0.
      
      * Revert "refine optimizer create accumulators (#50188)"
      
      This reverts commit 244e7546.
      
      * Revert "fix found_inf bug for custom optimizer (#50158)"
      
      This reverts commit 64573f9f.
      
      * Revert "refine amp scaler found_inf (#49864)"
      
      This reverts commit 382e9a06.
      
      * fix code format
      
      * fix conflict
      8a503522
  15. 22 2月, 2023 3 次提交
  16. 21 2月, 2023 3 次提交
  17. 20 2月, 2023 2 次提交
  18. 16 2月, 2023 1 次提交
  19. 15 2月, 2023 3 次提交
    • X
      align tool (#49865) · 4632ca13
      xu98bin 提交于
      * auto parallel align tool
      
      * modify function get_var's return
      
      * add save and load in align_tool
      
      * modify load function and save function
      
      * add finding different ops in align tool
      
      * full auto parallel align tool
      
      add test file for auto parallel align tool
      
      set timeout for test
      
      modify get_backward_tmp_var function
      
      add annotation for align tool
      
      modify test file
      
      modify code to restart CI
      
      remove timeout
      
      * set timeout
      4632ca13
    • W
      [fluid clean]clean fluid.transpiler API (#50375) · b08c91ab
      wangxiaoning 提交于
      * move ascend_transpiler
      
      * move transpiler.collective
      
      * remver checkport
      
      * fix
      
      * fix import
      
      * fix import
      
      * add init
      
      * fix
      
      * fix
      
      * fix
      b08c91ab
    • W
      [mv fluid] ps related (#50376) · 81113b53
      wangzhen38 提交于
      81113b53
  20. 13 2月, 2023 1 次提交
    • Y
      [Auto Parallel] Fix a bug of dist_scale (#50288) · 7f7e9320
      Yulong Ao 提交于
      * [Auto Parallel] Rename methods of ProcessMesh
      
      * [Auto Parallel] Impl the python process_mesh by the c++ one
      
      * [Auto Parallel] Add some minor modifications
      
      * [Auto Parallel] Rename some methods
      
      * [Auto Parallel] Remove unnecessary codes
      
      * [Auto Parallel] Add back some removed files
      
      * [Auto Parallel] Fix bugs
      
      * [Auto Parallel] Fix a bug
      
      * Update process_mesh.cc
      
      * [Auto Parallel] Merge dist attrs of Python into C++
      
      * [Auto Parallel] Add back deleted importing
      
      * [Auto Parallel] Add back removed unittest
      
      * [Auto Parallel] Remove type qualifiers of return types
      
      * [Auto Parallel] Fix some bugs
      
      * [Auto Parallel] Fix a bug of the quant pass
      
      * [Auto Parallel] Fix the code style
      
      * [Auto Parallel] Clear some fluid APIs
      
      * [Auto Parallel] Fix a bug of dist_scale
      7f7e9320