- 21 11月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Support many data types of several operators * SeqConv only support float/double * Revert adagrad
-
- 13 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* add adam op moment1_out = beta1 * moment1 + (1 − beta1) * grad moment2_out = beta2 * moment2 + (1 − beta2) * grad * grad moment1_hat = moment1_out / (1 - beta1^t) moment2_hat = moment2_out / (1 - beta2^t) param_out = param - learning_rate * moment1_hat / (sqrt(moment2_hat) + epsilon) * fix moment 2 * Adding the Adam optimization operator * Adding more tests for Adam op
-
- 07 8月, 2017 1 次提交
-
-
由 dongzhihong 提交于
-
- 04 8月, 2017 1 次提交
-
-
由 liaogang 提交于
-
- 31 7月, 2017 1 次提交
-
-
由 qijun 提交于
-
- 25 7月, 2017 1 次提交
-
-
由 Yu Yang 提交于
Make implement an operator less noisy.
-
- 19 7月, 2017 1 次提交
-
-
由 Qiao Longfei 提交于
* a simplest SGD op
-