- 15 6月, 2018 1 次提交
-
-
由 fengjiayi 提交于
-
- 11 6月, 2018 1 次提交
-
-
由 dzhwinter 提交于
* "add inplace attribute" * "register inplace attribute" * "change se-next model for memory-reuse" * "fix typo" * repick * fix merge conflict * "fix stupid error"
-
- 07 6月, 2018 2 次提交
-
-
由 Xin Pan 提交于
-
由 mozga-intel 提交于
* Add MKLDNN layout support in Paddle Add MKLDNN layout in Paddle so that MKLDNN friendly memory layout can be used in MKLDNN enabled OP kernel. Before this commit, NCHW is hardcode to be used in all MKLDNN op kernels. As a result, non-optimized execution path is selected in MKLDNN primitive which bring worse performance. Besides framework change, three MKLDNN OP kernels were updated for using new MKLDNN layout. They are conv/pool2d/batch_norm. Other MKLDNN OP kernels need be also updated in similar way to achieve best performance. * Add MKLDNN layout support in activation OP * Don't populate layout from input to output when kMKLDNN in * Refine pool mkldnn op kernel * MKLDNN layout * Remove the inferitance from tensor file * MKLDNN layout: refactoring * Remove additional #define to register new operator * Prepare mkldnn tests to work with layout
-
- 21 5月, 2018 2 次提交
-
-
由 Krzysztof Binias 提交于
-
由 Krzysztof Binias 提交于
-
- 08 5月, 2018 1 次提交
-
-
由 Yu Yang 提交于
Do not use ctor * Reduce line of codes. * We can use virtual function for Maker now. * The implementation does not care what maker holds, it is easier to refactor later.
-
- 21 4月, 2018 1 次提交
-
-
由 qiaolongfei 提交于
-
- 19 4月, 2018 1 次提交
-
-
由 Yang Yang(Tony) 提交于
* script to add semicolon * fix typo
-
- 18 4月, 2018 2 次提交
- 17 4月, 2018 2 次提交
- 16 4月, 2018 2 次提交
- 10 4月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 29 3月, 2018 1 次提交
-
-
由 chengduoZH 提交于
-
- 28 3月, 2018 1 次提交
-
-
由 chengduoZH 提交于
-
- 23 3月, 2018 3 次提交
-
-
由 Krzysztof Binias 提交于
-
由 Krzysztof Binias 提交于
-
由 Krzysztof Binias 提交于
-
- 21 3月, 2018 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 12 2月, 2018 1 次提交
-
-
由 qingqing01 提交于
-
- 10 2月, 2018 2 次提交
- 26 12月, 2017 2 次提交
- 20 12月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Move framework.proto to proto namespace * Fix compile * Fix compile * Fix Compile
-
- 12 12月, 2017 1 次提交
-
-
由 QI JUN 提交于
There are mainly following fixes: - take `DeviceContext` as the template parameter of math functors and OpKernel instead of `Place` - remove `eigen_device` interface in base class `DeviceContext` - remove `GetEigenDevice` interface in `ExecutionContext` and base class `DeviceContext` - remove unused `platform::EigenDeviceConverter` - rename `REGISTER_OP_GPU_KERNEL` to `REGISTER_OP_CUDA_KERNEL` - rename `USE_GPU_ONLY_OP` to `USE_CUDA_ONLY_OP`
-
- 07 12月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
-
- 06 12月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Using latex fraction syntax in sigmoid and logsigmoid op * Fixing the rendering of the latex equations in adamax operator
-
- 26 11月, 2017 1 次提交
-
-
由 dzhwinter 提交于
* "add floor, ceil, round op" * "reuse zero gradient" * "fix divide zero" * "fix numpy floor error"
-
- 21 11月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Support many data types of several operators * SeqConv only support float/double * Revert adagrad
-
- 05 11月, 2017 1 次提交
-
-
由 kexinzhao 提交于
* fix m_ops * fix activation op
-
- 03 11月, 2017 1 次提交
-
-
由 Kexin Zhao 提交于
-
- 27 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
* Simplize Gradient Check * Stash * Extract apply_backward_pass to backward.py Rename apply_backward_pass to append_backward_ops * Use graph API to check gradient * Fix ci * Fix CI * Fix backward for double precision * Stash * Fix CI * Fix ci * Ignore GRU test * Ignore xe op * Fix CI * Fix softmax with xe gradient The correct equation should be IG = OG * (d_softmax_with_xe()) * Fix typo * Fix merge error * Disable LRN
-
- 17 10月, 2017 1 次提交
-
-
由 Yu Yang 提交于
They are public now
-
- 13 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Adding Hard Sigmoid Activation * Adding a comment for slope to be only positive * Fixing grammatical mistake in comment
-
- 12 10月, 2017 1 次提交
-
-
由 Abhinav Arora 提交于
* Adding thresholded_relu op * Adding test for thresholded relu op
-
- 11 10月, 2017 1 次提交
-
-
由 kexinzhao 提交于
* implementing softplus * small fix * small fix * small fix * small fix
-