- 13 10月, 2017 2 次提交
-
-
由 Abhinav Arora 提交于
* add adam op moment1_out = beta1 * moment1 + (1 − beta1) * grad moment2_out = beta2 * moment2 + (1 − beta2) * grad * grad moment1_hat = moment1_out / (1 - beta1^t) moment2_hat = moment2_out / (1 - beta2^t) param_out = param - learning_rate * moment1_hat / (sqrt(moment2_hat) + epsilon) * fix moment 2 * Adding the Adam optimization operator * Adding more tests for Adam op
-
由 fengjiayi 提交于
-
- 12 10月, 2017 10 次提交
-
-
由 chengduoZH 提交于
-
由 chengduoZH 提交于
-
由 kexinzhao 提交于
* Implementing the DecayedAdagrad optimizer step operator * implementing DecayedAdagrad operator * remove file * small fix
-
由 武毅 提交于
* add cudnn_conv_op * WIP * update * update * fix grad check * use platform::memory * add support group for cudnn * update * follow comments * fix onlycpu build * update cuda define * follow comments * follow comments * merge with updates * fix compile error * follow comments * follow comments
-
由 chengduoZH 提交于
-
由 Yu Yang 提交于
-
由 chengduoZH 提交于
-
由 fengjiayi 提交于
-
由 Abhinav Arora 提交于
* Adding thresholded_relu op * Adding test for thresholded relu op
-
由 fengjiayi 提交于
-
- 11 10月, 2017 17 次提交
-
-
由 ranqiu 提交于
-
由 xzl 提交于
-
由 qiaolongfei 提交于
-
由 kexinzhao 提交于
* implementing softplus * small fix * small fix * small fix * small fix
-
由 fengjiayi 提交于
-
由 qiaolongfei 提交于
-
由 kavyasrinet 提交于
* Implemented the hardShrink activation * Fixing the unit test
-
由 fengjiayi 提交于
-
由 Siddharth Goyal 提交于
* Add numerically-stable logsigmoid activation * Add softshrink operator * Adjust relative tolerance for grad-check * Address review comments
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 Yu Yang 提交于
-
由 Yu Yang 提交于
-
由 fengjiayi 提交于
-
由 fengjiayi 提交于
-
由 Markus Kliegl 提交于
* conv_shift_op: initial implementation using Eigen Limitations: - both gradient outputs must be specified and are always computed - explicit for loops => could be optimized in various ways (e.g., different memory layout) * conv shift - gradient fixes fix case when not all output gradients desired * conv shift: minor cleanup * conv shift - more minor cleanup * conv shift: clean up & initial GPU implementation * fix rebase issue
-
- 10 10月, 2017 11 次提交
-
-
由 Yancey1989 提交于
-
由 fengjiayi 提交于
-
由 Yu Yang 提交于
1. InferShape should be a method for `OpDescBind` not `Operator`, since `OpDescBind` is a compile-time concept. 2. Pre-create operators and store them into a map. Make InferShape faster
-
由 fengjiayi 提交于
-
由 Abhinav Arora 提交于
-
由 Yu Yang 提交于
-
由 Yu Yang 提交于
-
由 Yu Yang 提交于
-
由 Abhinav Arora 提交于
* Implementing the Adamax optimizer step operator * Adding unit tests for adamax_op * Changing learning rate and time step to inputs from attributes * Changing learning rate and time step to input(tensors) * Making the Adamax operator conform to naming convention * Removing Tensor<float> from comments * Rectifying the Adamax implementation * Changing Unit Test values and adding comments * Changing Unit Test to test multiple steps
-
由 Yu Yang 提交于
-
由 kavyasrinet 提交于
-