- 18 10月, 2017 1 次提交
-
-
由 Markus Kliegl 提交于
* initial matmul operator Similar to np.matmul, but also has transpose_X and transpose_Y flags, and only supports tensors from rank 1 to 3 inclusive. For GPU, uses cublas?gemmStridedBatched. For CPU, uses cblas_?gemm_batch if available via MKL; otherwise a simple serial implementation that loops over the batch dimension is employed for now.
-
- 06 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Adding Adadelta optimization operator * Making inputs and outputs conform to naming convention * Removing type alias from header files * Fixing Adadelta documentation in comments * Addressing code review feedback
-
- 07 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 04 8月, 2017 1 次提交
-
-
由 liaogang 提交于
-
- 31 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 25 7月, 2017 1 次提交
-
-
由 Yu Yang 提交于
Make implement an operator less noisy.
-
- 19 7月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* a simplest SGD op
-