1. 16 3月, 2023 1 次提交
    • J
      [Auto Parallel Performance] Support BF16 Training (#51285) · 9ded5707
      JZ-LIANG 提交于
      * update env setting
      
      * update pass logic
      
      * dist op support bf16
      
      * backward cast update
      
      * update setting
      
      * update backward
      
      * revert amp pass
      
      * update fp16 backward logic
      
      * register c_embedding bf16
      
      * revert engine
      
      * add unitest
      
      * add unitest
      
      * update unitest
      
      * update cmake
      
      * update math
      
      * update math.py
      
      * update unitest
      
      * update unitest
      
      * revise unitest
      
      * revise unitest
      
      * update unitest
      
      * update unitest
      
      * update unitest
      9ded5707
  2. 04 1月, 2023 1 次提交
    • J
      [Auto Parallel-Performance] Sharding Comm Optimization (#48604) · 5592f8ad
      JZ-LIANG 提交于
      * remove deps and prior comm
      
      * grad comm fuse
      
      * add deps for amp&global norm
      
      * stage2 broadcast prior deps
      
      * stage2 grad overlap
      
      * stream_analyzer bugfix
      
      * overlap enable
      
      * dep op namescope
      
      * depend support multiple inputs
      
      * check finite deps
      
      * stage2 param comm overlap
      
      * Set kD2HStream
      
      * grad comm hierarchical
      
      * grad comm hierarchical
      
      * new unitest
      Co-authored-by: Nchenruibiao <chenruibiao@baidu.com>
      5592f8ad
  3. 29 12月, 2022 1 次提交
  4. 14 12月, 2022 1 次提交
    • Z
      [AutoParallel] recompute tuning (#48608) · 170a31f9
      zhaoyingli 提交于
      * [AutoParallel] recompute tuning
      
      * fix conflict
      
      * update comment
      
      * bug fix
      
      * update rc algo
      
      * tiny fix
      
      * fix clear process_group
      
      * remove comment
      
      * update segment print
      
      * fix import OpRole
      
      * adapt amp pass and grad_clip pass for opt_tuner
      
      * update tuning config
      
      * fix import
      
      * annotate recompute info on ops and upgrade recompute pass
      
      * add op_namescope for seed op
      
      * record reserved vars
      
      * fix recompute var's dist_attr
      
      * fix strategy unittest
      
      * adapt for fp16
      
      * update unittest
      
      * revert copy opt
      
      * update unittest
      
      * rename set_recompute_segments
      
      * fix unittest
      170a31f9
  5. 25 11月, 2022 1 次提交
  6. 08 11月, 2022 1 次提交
  7. 01 11月, 2022 1 次提交
    • N
      [CodeStyle][E711] use `is`/`is not` for comparison with `None` (#47452) · a35a4a53
      Nyakku Shigure 提交于
      * [CodeStyle][E711] use `is`/`is not` for comparison with `None`
      
      * `self.assertTrue($A is None)` -> `self.assertIsNone($A)`
      
      * `self.assertTrue($A is not None)` -> `self.assertIsNotNone($A)`
      
      * `self.assertFalse($A is None)` -> `self.assertIsNotNone($A)`
      
      * `self.assertEqual($A, None)` -> `self.assertIsNone($A)`
      
      * `self.assertNotEqual($A, None)` -> `self.assertIsNotNone($A)`
      a35a4a53
  8. 23 10月, 2022 1 次提交
  9. 12 10月, 2022 1 次提交
  10. 19 9月, 2022 1 次提交
  11. 17 9月, 2022 1 次提交
  12. 15 9月, 2022 1 次提交
    • Y
      [Auto Parallel] Improve the APIs (#45776) · b042a3b1
      Yulong Ao 提交于
      * [Auto Parallel] Use c++ dist attr in the completion process
      
      * [Auto Parallel] Add minor changes
      
      * [Auto Parallel] Use c++ dist attr in the completion process
      
      * [Auto Parallel] Add minor changes
      
      * [Auto Parallel] Add the serialization process for dist attrs
      
      * [Auto Parallel] Remove unnecessary comments
      
      * [Auto Parallel] Fix some bugs
      
      * [Auto Parallel] Fix the code style
      
      * [Auto Parallel] Remove unnecessary impls
      
      * [Auto Parallel] Fix the importing error
      
      * [Auto Parallel] Fix the copy from bugs of op dist attr
      
      * [Auto Parallel] Replace the use of constexpr if
      
      * [Auto Parallel] Redesign the shard_tensor, shard_op and ProcessMesh
      
      * [Auto Parallel] Change API of the completion unittest
      
      * [Auto Parallel] Fix the bug when set_attr an int
      
      * [Auto Parallel] Add the unittest for the serialization
      
      * [Auto Parallel] Add some unit tests
      
      * [Auto Paralle] Unify the strategy
      
      * [Auto Parallel] Improve the engine api
      
      * [Auto Parallel] Reset the changes made to the framework
      
      * [Auto Parallel] Change the engine unittest
      
      * [Auto Parallel] Update API of the completion and partitioner
      
      * [Auto Parallel] Update unit tests using engine api
      
      * update shard annotation
      
      * [Auto Parallel] Remove the modifications of other modules
      
      * [Auto Parallel] Add docs for APIs
      
      * add new strategy
      
      * [Auto Parallel] Replace the logger
      
      * [Auto Parallel] Restore the test_program.py
      
      * [Auto Parallel] Change the import rules
      
      * [Auto Parallel] Add the examples for Engine
      
      * [Auto Parallel] Do some minor changes
      
      * [Auto Parallel] Remove yaml dependency
      
      * [Auto Parallel] Fix the unittests
      
      * add valid after train
      
      * bug fix
      Co-authored-by: Nzhaoyingli <zhaoyingli@baidu.com>
      Co-authored-by: Ncaozhou <caozhou@radi.ac.cn>
      Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com>
      b042a3b1