1. 24 8月, 2021 1 次提交
    • Y
      Add auto completion module for auto parallel (#34813) · 93d862b0
      Yulong Ao 提交于
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * add dist
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update
      
      * update, test=develop
      
      * update, test=develop
      
      * update
      
      * update
      
      * delete unused proto
      
      * resotre op_desc
      
      * restore type_defs
      
      * update var_desc
      
      * remove dimss_mapping for proto_pybind
      
      * update interface.py
      
      * update framework.py
      
      * update
      
      * update
      
      * add auto_parallel dir
      
      * mv to paddle.distributed
      
      * add shard_xx api
      
      * add distributed attrs for var
      
      * add ut, test=develop
      
      * [WIP] Add the auto completion feature and related codes
      
      * [WIP] Improve the auto completion and related codes
      
      * [WIP] Make the auto completion to support data-parallel
      
      * [WIP] Make the completion support mp and dp+mp
      
      * [WIP] Refactor auto completion unit test for MLP
      
      * [WIP] Refactor the implementation of DistributedOperatorImpl
      
      * [WIP] Improve dims_mapping update rule and fix a bug
      
      * [WIP] Support auto completion for one transformer decoder layer
      
      * [WIP] Add a minor change
      
      * [WIP] Fix a bug within the uint test
      
      * Shard XShape tensor, add embedding completion and refactor code
      
      * Add the distributed_operators dir to setup.py.in
      
      * Improve the completion process and add the unittest for gpt
      
      * fix process_mesh ut
      
      * fix process_mesh ut
      
      * update
      
      * update, test=develop
      
      * Add support for automatically completing distributed attrs of special ops
      
      * update
      
      * update
      
      * update
      
      * fix doc sample codes, test=develop
      
      * improve coverage, test=develop
      
      * add static_mode check, test=develop
      
      * Model the cluster for cost model and physical mapping
      
      * update, test=develop
      
      * add set_placement, test=develop
      
      * Add the check to make sure the candidate tensors' size is great than zero
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update doc, test=develop
      
      * update, test=develop
      
      * Auto mark dist attrs annotated by user
      
      * update ndarray to nested list, test=develop
      
      * update, test=develop
      
      * Add auto-completion module for auto-parallel (based on PR#33804)
      
      * Remove unnecessary files
      
      * Remove unrelated files for the auto completion pr
      
      * Update the unit test to improve the coverage
      
      * Modify codes based on reviews
      
      * Minor changes for CI
      
      * Improve some codes based on new comments
      
      * Fix bugs caused by shallow copy in attributes.py
      * Imporve amend_distributed_attr_for_program in context.py
      * Other changes for weihang's comments
      Co-authored-by: Nsandyhouse <lilong12@baidu.com>
      93d862b0
  2. 23 8月, 2021 1 次提交
  3. 20 8月, 2021 2 次提交
  4. 18 8月, 2021 3 次提交
  5. 17 8月, 2021 2 次提交
    • C
      Copy boost optional to Paddle (#34780) · 9be41447
      chentianyu03 提交于
      * copy boost optional.hpp to paddle
      
      * copy boost optional.hpp to paddle
      
      * move directions
      
      * del fluid/utils
      
      * modify .hpp to .h
      
      * move directions
      
      * modify to paddle::optional
      
      * add modification description
      
      * format code stype for the files in paddle/utils
      
      * format code stype
      9be41447
    • Z
      Add some passes which can be applied to Program (#34730) · 8046e33d
      Zeng Jinle 提交于
      * add inplace passes and tests
      
      * update
      
      * fix use_cuda undefined
      fix compile error of op compat
      
      * add more ut
      
      * fix CPU CI error
      
      * check adam unique
      
      * fix mac/windows ci, improve coverage
      
      * fix ci error
      
      * follow weihang's comment
      
      * fix BlockDesc::MoveFrom
      
      * follow qiuliang's comment
      
      * update
      
      * follow huihuang's comments
      8046e33d
  6. 16 8月, 2021 2 次提交
  7. 13 8月, 2021 2 次提交
  8. 11 8月, 2021 4 次提交
  9. 10 8月, 2021 1 次提交
    • C
      copy boost/any.hpp to utils and replace boost::any with self defined any (#34613) · 12892929
      chentianyu03 提交于
      * add any.hpp to utils and replace boost::any with self defined paddle::any
      
      * add copy any.hpp to custom op depends
      
      * modify any.hpp include path
      
      * remove boost from setup.py.in
      
      * add copy any.hpp to custom op depends
      
      * move any.hpp to paddle/utils/ dirs
      
      * move any.h to extension/include direction
      
      * copy utils to right directions
      12892929
  10. 06 8月, 2021 3 次提交
  11. 05 8月, 2021 3 次提交
    • H
      New executor dev (#34407) · 012d12b5
      hong 提交于
      * first test version
      
      * add test exec;
      
      * add data transfer; test=develop
      
      * add new exec head;
      
      * add memcpy; test=develop
      
      * add python fetch
      
      * add new test
      
      * add graph node; test=develop
      
      * remove useless new executor test; test=develop
      
      * remove gperf dependency; test=develop
      
      * fix compile bugs; test=develop
      
      * remove useless code; test=develop
      
      * remove useless code; test=develop
      
      * add uni test; test=develop
      
      * polish code; test=develop
      
      * polish code; test=develop
      
      * add interpreter cmakefile; test=develop
      
      * remove useless code; test=develop
      012d12b5
    • C
      remove boost::algorithm::ends_with ,boost macro and boost::lexical_cast apis (#34310) · bb7b4c0c
      chentianyu03 提交于
      * replace boost::algorithm::ends_with with self define ends_with function
      
      * remove BOOST macro in certain operators
      
      * remove boost::lexical_cast
      
      * add test for string_helper
      
      * add more test case for string_helper
      
      * modify join_string func and test case
      
      * fix build_strategy_test failed bug
      
      * remove string_helper_test from parallel_UT_rule.py
      bb7b4c0c
  12. 04 8月, 2021 2 次提交
  13. 03 8月, 2021 2 次提交
  14. 02 8月, 2021 2 次提交
  15. 30 7月, 2021 3 次提交
    • H
      Revert of PR34452 (#34516) · 72a9c8ff
      Huihuang Zheng 提交于
      72a9c8ff
    • J
      Added reshape, reshape2, squeeze and squeeze2 BF16/FP32 FWD/BWD kernels (#34219) · 22c4c189
      jakpiase 提交于
      * test version of matmul_v2
      
      * added matmul_v2 grad kernel
      
      * minor changes
      
      * minor changes
      
      * minor change for CI approval
      
      * CI fix
      
      * CI fix
      
      * added squeeze and squeeze2 kernels
      
      * CI fix
      
      * CI fix
      
      * CI fix
      
      * disabled tests when compiled with cuda
      
      * added setting format_tag by strides
      
      * added sigmoid BF16 FWD/BWD and gelu BF16 BWD
      
      * changes after review
      
      * Revert "added sigmoid BF16 FWD/BWD and gelu BF16 BWD"
      
      This reverts commit 6e3f76720b545abfcff9f6052b46b73a1e745cae.
      
      * Revert "Merge branch 'matmul_v2_grad' into squeeze2_op"
      
      This reverts commit 06fcf67843a4a7884eccdf67a02a03575e1d4cb8, reversing
      changes made to 6e3f76720b545abfcff9f6052b46b73a1e745cae.
      
      * minor change
      
      * added reshape1/2 kernels
      
      * moved some functions into private block
      
      * CI fix
      
      * CI fix
      
      * CI fix
      22c4c189
    • W
      add trainer desc config to distributed strategy (#34457) · e6aacd1e
      wangguanqun 提交于
      * add trainer desc config to distributed strategy
      
      * code style modified
      e6aacd1e
  16. 29 7月, 2021 5 次提交
  17. 28 7月, 2021 2 次提交