1. 12 12月, 2017 1 次提交
  2. 16 11月, 2017 1 次提交
    • Y
      feature/while_grad_op (#5554) · 18f0c40a
      Yang Yang(Tony) 提交于
      * first commit
      
      * Python API for while op
      
      * Python Unittest for simple while_op forward
      
      * fix out to be list
      
      * Fix UT
      
      * VarType
      
      * Fix several bugs
      
      * Fix bug
      
      * Fix bug
      
      * Fix Bug
      
      * Fix bug
      
      * Fix unittest
      
      * Remove debug log
      
      * Add comments
      
      * add PADDLE_ENFORCE
      
      * while_grad_op first commit
      
      * Add `BlockDescBind::FindRecursiveOrCreateVar()` and fix bugs
      
      * not sure how to setdim of while outputs
      
      * push for test
      
      * add executor vlog
      
      * fix bug of while_op cond
      
      * Several enhancement for code
      
      1. Backward always infer shape & infer var type. Since there are RENAME
      variables will be created when creating backward operator, but their
      shape & var types are not inferenced.
      2. Never use SomePtr-> directly, since every pointer could be nullptr if
      it is a function return value. Add `detail::Ref` to cast pointer to
      reference safely.
      3. Enhance error message for backward.
      4. Infer data type of variable in `sum` and `tensor_write`
      
      * Fix bugs of while_op gradient
      
      * Fix several bugs of while_op grad
      
      * fix fill zeros like
      
      * fix 3 >= 3
      
      * fix place holder shouldn't be null
      
      * fail on sum op
      
      * Fix SumOp of TensorList
      
      * clean up
      
      * pass while test
      
      * fix test_array_write_read
      
      * pass sum op
      
      * Support int/int64 for fill_constant_batch_size_like
      
      * Fix compile
      18f0c40a
  3. 14 11月, 2017 1 次提交
  4. 11 11月, 2017 2 次提交
  5. 08 11月, 2017 2 次提交
  6. 26 10月, 2017 1 次提交
  7. 18 10月, 2017 1 次提交
    • M
      MatMul operator (#4856) · 16489827
      Markus Kliegl 提交于
      * initial matmul operator
      
      Similar to np.matmul, but also has transpose_X and transpose_Y flags,
      and only supports tensors from rank 1 to 3 inclusive.
      
      For GPU, uses cublas?gemmStridedBatched. For CPU, uses
      cblas_?gemm_batch if available via MKL; otherwise a simple serial
      implementation that loops over the batch dimension is employed for now.
      16489827
  8. 16 10月, 2017 1 次提交
  9. 15 10月, 2017 1 次提交
  10. 14 10月, 2017 4 次提交
  11. 21 9月, 2017 1 次提交
  12. 19 9月, 2017 1 次提交
  13. 22 8月, 2017 2 次提交
  14. 21 8月, 2017 3 次提交
  15. 19 8月, 2017 1 次提交
  16. 14 8月, 2017 2 次提交
  17. 11 8月, 2017 2 次提交
  18. 10 8月, 2017 1 次提交
  19. 07 8月, 2017 2 次提交
  20. 03 8月, 2017 2 次提交