- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 11 10月, 2017 1 次提交
-
-
由 guosheng 提交于
-
- 06 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Adding Adadelta optimization operator * Making inputs and outputs conform to naming convention * Removing type alias from header files * Fixing Adadelta documentation in comments * Addressing code review feedback
-
- 07 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 04 8月, 2017 1 次提交
-
-
由 liaogang 提交于
-
- 31 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 25 7月, 2017 1 次提交
-
-
由 Yu Yang 提交于
Make implement an operator less noisy.
-
- 19 7月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* a simplest SGD op
-