1. 24 11月, 2022 1 次提交
  2. 24 10月, 2022 2 次提交
  3. 19 10月, 2022 1 次提交
    • Z
      [Cherry-Pick][AutoParallel] auto_parallel cherry-pick to release2.4 (#47145) · 90b31790
      zhaoyingli 提交于
      * [Auto Parallel] Make Engine class callable (#46416)
      
      * [Auto Parallel] Imporve the user-defined fetches and logging
      
      * [Auto Parallel] Make Engine class callable
      
      * [Auto Parallel] Update the data loading of tuner
      
      * Print IPS in auto parallel Engine (#46554)
      
      * [AutoParallel] fix dist_split (#46505)
      
      * [AutoParallel] fix dist_split
      
      * add unittest
      
      * update cmakelist
      
      * [AutoParallel] fix sharding (#46572)
      
      * [AutoParallel] fix process_mesh (#46583)
      
      * [AutoParallel] fix reshard when train with eval (#46605)
      
      * [AutoParallel] fix reshard when train with eval
      
      * fix mppp
      
      * [AutoParallel] fix amp when predict (#46637)
      
      * [Auto Parallel]Update comp cost and completion for gpt auto search (#46387)
      
      * update comp cost and completion for gpt auto search
      
      * add unittest
      
      * [Auto Parallel] Fix bugs caused by the inconsistent outputs of Engine API (#46633)
      
      * [Auto Parallel] Unify the logger and outputs of Engine API
      
      * [Auto Parallel] Fix the bugs of to_static
      
      * [Auto Parallel] Adjust the test_to_static.py
      
      * [Auto Parallel] Improve the fine-grained APIs (#46552)
      
      * [Auto Parallel] Suppport different dataloaders
      
      * [Auto Parallel] Add num_shards config for dataset
      
      * [Auto Parallel] Unify the logger and outputs of Engine API
      
      * [Auto Parallel] Fix the bugs of to_static
      
      * [Auto Parallel] Adjust the test_to_static.py
      
      * [Auto Parallel] Add the prepare API and replace __call__ with run
      
      * [Auto Parallel] Improve the private implementations of Engine
      
      * [Auto Parallel] Set capacity of dataloader for opt tuning
      
      * [Auto Parallel] [WIP] Change the fine-grained API
      
      * [Auto Parallel] Improve APIs to support different user cases
      
      * [Auto Parallel] Add removed config
      
      * [Auto Parallel] Add imports
      
      * [Auto Parallel] Fix bugs for to_static
      
      * [Auto Parallel] Remove unnecessary imports
      
      * bugfix (#46921)
      
      * [Auto Parallel] Fix the bug for None labels (#46987)
      
      * [AutoParallel] adapt for gpt-gen (#46771)
      
      * for gpt-gen
      
      * fix reshard
      
      * adapt assign and shape op
      
      * add dist_assign & unittest
      
      * add conditional block unittest
      
      * rename unittest
      
      * [Auto Parallel] Fix the bug of completion (#47056)
      
      * [Auto Parallel] Fix the bug for None labels
      
      * [Auto Parallel] Fix the completion bug
      
      * [AutoParallel] add callbacks (#47014)
      
      * [AutoParallel] add callbacks
      
      * fix unittest
      
      * fix dist_context
      
      * fix engine
      
      * fix cmakelist
      
      * fix unittest's returns
      
      * fix cmakelist
      
      * [Auto Parallel] Add cost interface (#47043)
      
      * add cost interface
      
      * update inferface and add unittest
      
      * update unittest
      
      * update inferface
      
      * [Auto Parallel]Add parallel tuner (#46189)
      
      * add parallel tuner
      
      * add unittest
      
      * fix unittest
      
      * set timeout of unittest
      
      * set unittest timeout
      
      * fix auto_mode setting
      
      * update unittest
      
      * sync from develop and update unittest
      
      * remove unused import
      
      * update unittest
      
      * update cmakelist
      
      * add unittests
      Co-authored-by: NYulong Ao <aoyulong@baidu.com>
      Co-authored-by: NRuibiao Chen <chenruibiao@baidu.com>
      Co-authored-by: Ncaozhou <48191911+Caozhou1995@users.noreply.github.com>
      Co-authored-by: NJZ-LIANG <jianzhongliang10@gmail.com>
      90b31790
  4. 11 7月, 2022 1 次提交
  5. 23 6月, 2022 1 次提交
    • N
      fix paddle.vision.datasets.* en docs (#43649) · 1fca8f33
      Nyakku Shigure 提交于
      * rewrite all code examples, test=document_fix
      
      * refine arguments, test=document_fix
      
      * fix desc format error, test=document_fix
      
      * capitalize the first letter, test=document_fix
      
      * refine api desc, test=document_fix
      
      * fix wrong COPY-FROM label in Model docs, test=document_fix
      
      * refine returns, test=document_fix
      
      * refine returns, test=document_fix
      
      * add a blank line in code block, test=document_fix
      1fca8f33
  6. 17 6月, 2022 1 次提交
    • N
      fix paddle.Model en docs (#43537) · 4c3969fa
      Nyakku Shigure 提交于
      * add copy-from label for code examples, test=document_fix
      
      * refine docs, test=document_fix
      
      * add some output for code example, test=document_fix
      
      * add `optional`, test=document_fix
      
      * add missing parameters, test=document_fix
      
      * add missing links for `ProgBarLogger` and `ModelCheckpoint`, test=document_fix
      
      * update eval_batch example, test=document_fix
      
      * fix typos in stack_outputs, test=document_fix
      
      * np.random -> paddle.random, test=document_fix
      4c3969fa
  7. 05 6月, 2022 1 次提交
    • S
      【code format check upgrade】 step2:yapf (#42944) · a072fca8
      Sing_chan 提交于
      * use yapf to format all python file
      
      * yapf exclude two unittests file for they rely on writing and reading file, and format will break them
      
      * disable diff_py_file because too many diff files cause command following failed
      a072fca8
  8. 13 5月, 2022 1 次提交
  9. 12 5月, 2022 1 次提交
  10. 25 3月, 2022 1 次提交
    • J
      Refactor Dygraph Flags (#40786) · 3085d5e4
      Jiabin Yang 提交于
      * refactor eager flags
      
      * fix flags error when we switch from eager to dygraph
      
      * fix ci problem
      
      * fix ci
      
      * fix ci
      
      * merge develop and fix code style
      
      * merge develop and fix code style
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * fix op test error
      
      * merge develop
      3085d5e4
  11. 21 3月, 2022 1 次提交
    • H
      Merge some test bug (#40543) · 56c43ccd
      hong 提交于
      * switch eager mode and change it
      
      * set default is eager
      
      * set default is eager
      
      * fix error; test=develop
      
      * fix some error; test=develop
      
      * update
      
      * upd
      
      * update code; test=develop
      
      * update
      
      * fix some bug; test=develop
      
      * fix bug; test=develop
      
      * fix bug; test=develop
      
      * fix bug; test=develop
      
      * fix error; test=develop
      
      * format; test=develop
      Co-authored-by: NJiabinYang <360788950@qq.com>
      56c43ccd
  12. 22 10月, 2021 1 次提交
    • L
      [hapi] support dygraph amp O2 (#36441) · 08248db0
      Leo Chen 提交于
      * [hapi] support dygrapg amp O2
      
      * fix problem of static pure fp16 in hapi
      
      * fix bug
      
      * fix format
      
      * fix ut
      
      * follow comments
      
      * update ut
      
      * update amp save/load
      
      * fix ut
      
      * refine code format
      08248db0
  13. 29 7月, 2021 1 次提交
  14. 23 7月, 2021 1 次提交
  15. 22 7月, 2021 1 次提交
  16. 08 7月, 2021 1 次提交
  17. 28 6月, 2021 8 次提交
  18. 21 6月, 2021 1 次提交
  19. 11 6月, 2021 1 次提交
  20. 09 6月, 2021 1 次提交
  21. 07 6月, 2021 1 次提交
  22. 29 4月, 2021 1 次提交
  23. 26 4月, 2021 2 次提交
  24. 23 4月, 2021 1 次提交
  25. 21 4月, 2021 1 次提交
    • Z
      【NPU】Merge NPU ccl code (#32381) · c3158527
      zhang wenhui 提交于
      * add allreduce and broadcast without test (#31024)
      
      add allreduce and broadcast without test
      
      * Refactor HCCLCommContext to be compatible with Paddle (#31359)
      
      Refactor HCCLCommContext to be compatible with Paddle (#31359)
      
      * [NPU] add npu kernel for communication op (#31437)
      
      * add allreduce and broadcast without test
      
      * add c_broadcast_test case
      
      * build c_comm_init and c_create_group operators
      
      * make the whole thing compile
      
      * add broadcast and init op test case but run failed
      
      * make unit test compile
      
      * fix broadcast test bug and change into hcom for ccl
      
      * change c_comm_init and c_create_group ops accordingly
      
      * make tests compile
      
      * transfer code to 27
      
      * compiled successfully in 28, but run failed
      
      * test broadcast in 28, but failed
      
      * make hcom primitives work
      
      * change hccl data type for base.h
      
      * fix broadcast bug
      
      * make attributes work
      
      * fix group name bug
      
      * add allreduce but test failed
      
      * allreduce bug for qiuliang
      
      * allreduce finished
      
      * add allgather and reducescatter
      
      * merge all op code
      
      * add allgather test
      
      * finish run all ccl op test exclude send/recv
      
      * all all op and test exclude send/recv
      
      * send_v2_npu.cc recv_v2_npiu.cc compiled
      
      * fix ccl core dump bug and test allgather, reducescatter, broadcast op
      
      * fix allreduce bug just for test
      
      * hcom send&recv test pass, without hcom_destroy
      
      * for qiuliang test
      
      * Ascend Send&Recv Test Pass
      
      * all op (ex send/recv) ok
      
      * fix bug
      
      * merge all ccl op
      
      * style merge to PaddlePaddle
      
      * merge style
      
      * new merge style
      
      * merge style 2
      
      * insert an empty at the end
      
      * disable ctest for hcom to pass ci
      Co-authored-by: Nvoid-main <voidmain1313113@gmail.com>
      Co-authored-by: Nf2hkop <f2huestc@outlook.com>
      
      * Add auto-increasing tag id for Hcom OPs (#31702)
      
      * add c_reduce_sum op (#31793)
      
      add c_reduce_sum op
      
      * update Ascendrc hccl to 20.3 (#32126)
      
      update Ascendrc hccl to 20.3 (#32126)
      
      * fix merge code
      
      * change cmake.txt1
      
      * [NPU] Support npu kernel for c sync stream op (#31386)
      
      * sync stream npu op
      
      * add with_ascend_acl
      
      * update c++ unittest
      
      * compile all failed
      
      * try to pre commit
      
      * after pre commit
      
      * merge&compile&test hccl successfully!
      
      * fix code style
      
      * fix code style
      
      * fix bugs about hccl
      
      * fix some bugs
      
      * fix code style
      
      * fix style
      
      * fix style
      
      * fix
      
      * fixed
      
      * merge develop
      Co-authored-by: Nlw921014 <liuwei921014@yeah.net>
      Co-authored-by: NVoid Main <voidmain1313113@gmail.com>
      Co-authored-by: Nf2hkop <f2huestc@outlook.com>
      Co-authored-by: Nxiayanming <41795079@qq.com>
      c3158527
  26. 15 4月, 2021 1 次提交
    • J
      Update hapi to support AMP (#31417) · fabdb43c
      Jiaqi Liu 提交于
      * make hapi support amp, and add unittest
      
      * make unittest only support GPU
      
      * update parameters for amp in hapi.Model
      
      * update hapi.Model.prepare interface, and update unittest
      
      * fix test_model.py unittest bug
      
      * add grad clear in dygraph
      
      * use_fp16_guard defaults to True, which could avoid nan
      
      * add input check, and add internal doc link to low level api
      
      * update doc, and decrease the sample num of dataset to avoid timeout
      
      * make hapi amp param  support str 'O1' or 'O2'
      
      * resume calling , modify the code of the check part
      
      * upgrade the usage of Fleet API, and disable 'pure_fp16' param
      fabdb43c
  27. 11 1月, 2021 1 次提交
  28. 07 1月, 2021 1 次提交
  29. 27 11月, 2020 2 次提交
    • S
      Support dynamic graph distributed (#28997) · e2d01eb6
      ShenLiang 提交于
      * add reducer
      
      * refine envent for memorycopy
      
      * add concat&split for allreduce
      
      * apply concat & split for fuse tensor
      
      * fix nccl dep
      
      * fix the untest, compile problem and ddp initialize problem
      
      * fix untest for mac & add some comments & solve the repeated param in sublayers
      
      * fix untest for windows & fix document
      e2d01eb6
    • L
      Enhance logger callback for benchmark (#29106) · 9f53f3d0
      LielinJiang 提交于
      * enhance logger callback for benchmark
      9f53f3d0
  30. 25 11月, 2020 1 次提交