- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 2 次提交
- 11 11月, 2017 1 次提交
-
-
由 dangqingqing 提交于
-
- 18 10月, 2017 1 次提交
-
-
由 Markus Kliegl 提交于
* initial matmul operator Similar to np.matmul, but also has transpose_X and transpose_Y flags, and only supports tensors from rank 1 to 3 inclusive. For GPU, uses cublas?gemmStridedBatched. For CPU, uses cblas_?gemm_batch if available via MKL; otherwise a simple serial implementation that loops over the batch dimension is employed for now.
-
- 10 8月, 2017 3 次提交
- 13 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 11 7月, 2017 2 次提交
- 04 7月, 2017 3 次提交
- 03 7月, 2017 2 次提交