- 08 5月, 2018 1 次提交
-
-
由 Yu Yang 提交于
Do not use ctor * Reduce line of codes. * We can use virtual function for Maker now. * The implementation does not care what maker holds, it is easier to refactor later.
-
- 06 5月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "optimizer op support float64" * "fix ci" * "fix ftrl op"
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 2 次提交
- 20 12月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Move framework.proto to proto namespace * Fix compile * Fix compile * Fix Compile
-
- 12 12月, 2017 2 次提交
-
-
由 QI JUN 提交于
There are mainly following fixes: - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place` - remove `eigen_device` interface in base class `DeviceContext` - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext` - remove unused `platform::EigenDeviceConverter` - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL` - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
-
由 kavyasrinet 提交于
* Updating the Latex equation for Adagrad * Fixing Latex euqations for adadelta, adam and adamax
-
- 06 12月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Using latex fraction syntax in sigmoid and logsigmoid op * Fixing the rendering of the latex equations in adamax operator
-
- 21 11月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Support many data types of several operators * SeqConv only support float/double * Revert adagrad
-
- 05 11月, 2017 1 次提交
-
-
由 kavyasrinet 提交于
* Adding the doc format for AdaDelta * Updating the documentation for Adagrad, Adam and Adamax * Updating the auc op * Fix review comments * Updating doc for Batch Norm * Updating the cast op * Updating the clip op * Fixing review comment * Fixing review comment: * Small change to restart PR_CI
-
- 20 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
-
- 17 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
They are public now
-
- 10 10月, 2017 2 次提交
-
-
由 Yu Yang 提交于
-
由 Abhinav Arora 提交于
* Implementing the Adamax optimizer step operator * Adding unit tests for adamax_op * Changing learning rate and time step to inputs from attributes * Changing learning rate and time step to input(tensors) * Making the Adamax operator conform to naming convention * Removing Tensor<float> from comments * Rectifying the Adamax implementation * Changing Unit Test values and adding comments * Changing Unit Test to test multiple steps
-