1. 10 2月, 2018 2 次提交
  2. 03 2月, 2018 1 次提交
  3. 12 12月, 2017 2 次提交
    • Q
      Refine device context (#6433) · 61ec0b95
      QI JUN 提交于
      There are mainly following fixes:
      
      - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place`
      - remove `eigen_device` interface in base class  `DeviceContext`
      - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext`
      - remove unused `platform::EigenDeviceConverter`
      - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL`
      - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
      61ec0b95
    • T
      unify MKL macro definition · 69b44f2f
      tensor-tang 提交于
      69b44f2f
  4. 26 11月, 2017 1 次提交
    • D
      Feature/copytensor (#5455) · 45062fe5
      dzhwinter 提交于
      * "make global tensor function independently"
      
      * "replace functor"
      
      * "fix inline template error"
      
      * "fix tensor array with CopyFrom"
      
      * "fix other case use CopyFrom"
      
      * "move the op interface hardly"
      
      * "fix operators"
      
      * "fix typo"
      
      * "delete dynamic recurrent rnn and fix gru_unit in debugmode"
      
      * "fix unique_ptr copy"
      
      * "fix cuda copy"
      
      * "fix namespace error"
      
      * "removed nccl python test"
      
      * "fix include error"
      
      * "fix typo"
      
      * fix copy util test
      45062fe5
  5. 14 11月, 2017 1 次提交
  6. 11 11月, 2017 1 次提交
  7. 09 11月, 2017 1 次提交
  8. 08 11月, 2017 1 次提交
  9. 26 10月, 2017 1 次提交
  10. 18 10月, 2017 1 次提交
    • M
      MatMul operator (#4856) · 16489827
      Markus Kliegl 提交于
      * initial matmul operator
      
      Similar to np.matmul, but also has transpose_X and transpose_Y flags,
      and only supports tensors from rank 1 to 3 inclusive.
      
      For GPU, uses cublas?gemmStridedBatched. For CPU, uses
      cblas_?gemm_batch if available via MKL; otherwise a simple serial
      implementation that loops over the batch dimension is employed for now.
      16489827
  11. 16 10月, 2017 1 次提交
  12. 14 10月, 2017 2 次提交
  13. 29 9月, 2017 1 次提交
  14. 21 9月, 2017 1 次提交
  15. 19 9月, 2017 1 次提交
  16. 22 8月, 2017 2 次提交
  17. 21 8月, 2017 3 次提交
  18. 19 8月, 2017 1 次提交
  19. 14 8月, 2017 3 次提交
  20. 10 8月, 2017 2 次提交
  21. 09 8月, 2017 1 次提交
  22. 07 8月, 2017 1 次提交
  23. 03 8月, 2017 2 次提交