1. 04 11月, 2022 1 次提交
  2. 03 11月, 2022 1 次提交
  3. 01 11月, 2022 1 次提交
  4. 29 10月, 2022 1 次提交
  5. 24 10月, 2022 3 次提交
  6. 21 10月, 2022 1 次提交
  7. 19 10月, 2022 1 次提交
    • G
      Add enable_partial_send_recv switch in pipeline_configs (#46992) (#47083) · 1d015f12
      Ghost Screaming 提交于
      * Fix bug of reduce_sum op. When input.numel() > INT32_MAX, its result
      is wrong.
      
      * Support allow_partial switch, which can be configure in
      pipeline_configs. If sent tensor are not the same from
      different hosts, they shouldn't been sent partially and
      then concated as a whole tensor.
      
      * Change name allow_partial to enable_partial_send_recv.
      
      * Add global variable _enable_partial_send_recv
      1d015f12
  8. 18 10月, 2022 2 次提交
    • Y
      Cherry pick for sharding (#47061) · 5b642140
      Yuang Liu 提交于
      * [dygraph sharding] Overlap the reduce and the caculation for sharding stage 2. (#46495)
      
      * [dygraph sharding stage 2] sharding broadcast overlap (#46656)
      
      * Multi groups for broadcast of sharding stage 2 (#46894)
      5b642140
    • H
      [cherry-pick] Fix perf issues of mp/pp/fuse in eager mode (#47071) · b84edd90
      Haohongxiang 提交于
      * [Dygraph] Fix performance of pp+mp by using send/recv_calc_stream instead of send/recv (#46116)
      
      * [Dygraph] Fix Perf of FusedFeedForward and FusedAttention with AllReduce (#46780)
      
      * update
      b84edd90
  9. 17 10月, 2022 1 次提交
    • W
      [Cherry-pick] Collective communication APIs (#46922) · 5fba2a98
      Wen Sun 提交于
      * Support both use_calc_stream and sync_op in send recv APIs (#46023)
      
      * Support both use_calc_stream and sync_op in allgather API (#46295)
      
      * Support both use_calc_stream and sync_op in collective communication API (#46761)
      
      * Move group and all reduce from collective to communication (#45848)
      
      * Completes bfloat16 dtype for collective api in eager mode (#45844)
      
      * Fix collective APIs cannot be recognized when building docs (#46962)
      Co-authored-by: NLiYuRio <63526175+LiYuRio@users.noreply.github.com>
      5fba2a98
  10. 11 10月, 2022 1 次提交
    • Y
      Cherry pick for dygraph pp (#46876) · 9cc3f69f
      Yuang Liu 提交于
      * bug fix for virtual pipeline parallel (#45922)
      
      * dont wait for send op under dygraph pp (#46209)
      
      * [interleave pp] sync recv for 1f1b (#46399)
      
      * [dygraph pp] all sync for allgather partial (#46483)
      9cc3f69f
  11. 27 9月, 2022 1 次提交
  12. 22 9月, 2022 2 次提交
  13. 20 9月, 2022 2 次提交
  14. 19 9月, 2022 3 次提交
  15. 09 9月, 2022 1 次提交
  16. 07 9月, 2022 2 次提交
  17. 06 9月, 2022 1 次提交
  18. 02 9月, 2022 1 次提交
  19. 01 9月, 2022 1 次提交
  20. 26 8月, 2022 3 次提交
  21. 23 8月, 2022 2 次提交
  22. 16 8月, 2022 1 次提交
  23. 15 8月, 2022 1 次提交
    • W
      refactor fleet. (#44833) · 8636d2a2
      wuhuachaocoding 提交于
      * refactor fleet.
      
      * refact fleet.py.
      
      * update fleet/__init__.py.
      
      * update fleet.py
      
      * update code style.
      
      * update fleet
      
      * update fleet
      
      * update fleet
      
      * update fleet
      
      * update model.py
      
      * update fleet.
      
      * update __init__.py
      
      * update fleet.
      
      * update fleet.
      
      * update fleet
      
      * update fleet
      
      * update fleet
      
      * update fleet.
      
      * update optimizer.py
      
      * update optimizer
      
      * update fleet.py
      
      * update scaler.py
      
      * update setup.py.in
      8636d2a2
  24. 13 8月, 2022 1 次提交
    • Z
      fl-ps: support split sparse params in local & remote (#44864) · 3f5c405f
      ziyoujiyi 提交于
      * back fl
      
      * delete ssl cert
      
      * .
      
      * make warning
      
      * .
      
      * unittest paral degree
      
      * solve unittest
      
      * heter & multi cloud commm ready
      
      * .
      
      * .
      
      * fl-ps v1.0
      
      * .
      
      * support N + N mode
      
      * .
      
      * .
      
      * .
      
      * .
      
      * delete print
      
      * .
      
      * .
      
      * .
      
      * .
      
      * fix bug
      
      * .
      
      * .
      
      * fl-ps with coordinator ready
      
      * merge dev
      
      * update message parse only
      
      * update fl client scheduler
      
      * fix bug
      
      * update multithreads sync
      
      * fix ci errors
      
      * update role_maker.py
      
      * update role_maker.py
      
      * fix ci error: windows py import error
      
      * fix ci error: windows py import error
      
      * fix windows ci pylib import error
      
      * add dump fields & params
      
      * try to fix windows import fleet error
      
      * fix ps FLAGS error
      
      * fix logging risk
      
      * fix logging possible risk
      
      * write trainer_desc file
      
      * support split sparse params in local & remote
      
      * fix import paddle.fluid.core.PSGPU
      
      * fix import paddle.fluid.core.PSGPU
      
      * add remote_sparse & local_sparse config
      
      * fix unittest
      
      * fix test_dist_fleet_geo table error
      
      * fix PADDLE_ENFORCE error
      
      * fix other's pr conflict
      3f5c405f
  25. 12 8月, 2022 1 次提交
  26. 10 8月, 2022 1 次提交
  27. 09 8月, 2022 2 次提交
    • Z
      refine save/load interface for distributed cpups (#44862) · 7b29c89b
      zhaocaibei123 提交于
      * save load
      
      * save load
      
      * add unittest
      
      * first commit
      
      * second commit
      
      * third commit
      
      * remove SaveLocalFS in memory sparse table
      
      * save dense param
      
      * update
      
      * push slot
      
      * fix push show clk: int -> float
      
      * add unittest
      
      * fix sample
      
      * unittest
      
      * add AsExtra for op
      
      * unittest
      
      * modify fs.py
      
      * modify fs.py
      
      * fix some bugs
      
      * add dataset hdfs config
      
      * local change
      
      * dataset use differenct hadoop ugi/fs_name
      
      * add
      
      * fix conflict
      
      * fix
      
      * remove logs
      
      * code style
      
      * fix
      
      * code style
      
      * code style
      
      * fix
      
      * code style
      
      * save_dense_param
      
      * fix
      
      * fix
      
      * fix
      
      * fix
      
      * change momentum in dense optimzer
      
      * fix
      
      * fix
      
      * change fluid => paddle.static
      
      * remove some unuseful code
      Co-authored-by: Nesythan <esythan@126.com>
      7b29c89b
    • Y
      [model parallel] enable mp to use fused linear (#44968) · e84250e8
      Yuang Liu 提交于
      e84250e8
  28. 08 8月, 2022 1 次提交