1. 04 11月, 2020 2 次提交
    • L
      Add broadcast_shape api (#28257) · 8b2436a7
      Leo Chen 提交于
      * add broadcast_shape api
      
      * add ut
      
      * follow comments
      
      * add example code, test=dodument_fix
      
      * update example code, test=document_fix
      8b2436a7
    • enhance the op_version_registry, test=develop (#28347) · 21a63f6f
      石晓伟 提交于
      * enhance the op_version_registry, test=develop
      
      * add unittests, test=develop
      
      * enhance the op_version_registry, test=develop
      
      * fix bugs, test=develop
      
      * revert pybind_boost_headers.h, test=develop
      
      * fix a attribute bug, test=develop
      21a63f6f
  2. 03 11月, 2020 4 次提交
  3. 02 11月, 2020 4 次提交
  4. 30 10月, 2020 4 次提交
  5. 29 10月, 2020 3 次提交
  6. 28 10月, 2020 8 次提交
  7. 27 10月, 2020 5 次提交
  8. 26 10月, 2020 9 次提交
  9. 23 10月, 2020 1 次提交
    • H
      Fix test_parallel_executor_test_while_train Random Failure by Decreasing GPU Usage (#28213) · a1e7fd4a
      Huihuang Zheng 提交于
      Recently, test_parallel_executor_test_while_train randomly failed on CI. On all CI logs, it showed NCCL initialization failed or cusolver initialization failed. I found online that those failure is usually caused by GPU shortage. Those API calls CUDA APIs directly so it shouldn't be the problem of allocator. It may be somewhere in PaddlePaddle increases GPU usage.
      
      However, I run this test for 1000 times on my machine and the CI machine, either of them can reproduce the random failure. Maybe there is something related to the environment only happened in test env.
      
      To verify my assumption that somewhere in PaddlePaddle increases GPU usage and also fix this CI, I decreased the batch_size to see whether the random failure disappears in test env.
      a1e7fd4a