- 14 9月, 2019 1 次提交
-
-
由 Adam 提交于
test=develop
-
- 04 9月, 2019 1 次提交
-
-
由 Adam 提交于
test=develop
-
- 03 9月, 2019 1 次提交
-
-
由 Adam 提交于
* using MKLDNNMemoryFormat = mkldnn::memory::format changes test=develop * PADDLE_ENFORCE update test=develop
-
- 29 8月, 2019 1 次提交
-
-
由 lidanqing 提交于
test=develop
-
- 15 8月, 2019 1 次提交
-
-
由 Adam 提交于
test=develop
-
- 10 6月, 2019 1 次提交
-
-
由 Zeng Jinle 提交于
* remove attribute in Allocator::Allocate, test=develop * fix travis ci error, test=develop
-
- 22 5月, 2019 1 次提交
-
-
由 guomingz 提交于
* Relu6 is the bottleneck op for Mobilenet-v2. As the mkldnn supports the conv/relu6 fusion, we implement it fusion via cpass way. Due to the int8 enabling for this fusion will be supported in MKLDNN v0.20, so this PR is focused on the fp32 optimization. Below table shows the benchmark(FPS) which measured on skx-8180(28 cores) Batch size | with fusion | without fusion -- | -- | -- 1 | 214.7 | 53.4 50 | 1219.727 | 137.280 test=develop * Fix the format issue test=develop * Add the missing nolint comments. test=develop * Fix the typos. test=develop * Register the conv_brelu_mkldnn_fuse_pass for the MKLDNN engine. test=develop * Adjust the indentation. test=develop * Add the test_conv_brelu_mkldnn_fuse_pass case. test=develop * Slightly update the code per Baidu comments. Let the parameter definition embedded into the code. That's will make the code easy to understand. test=develop
-
- 16 4月, 2019 1 次提交
-
-
由 Jacek Czaja 提交于
* - Reuse of conv PD - conv transpose pd reused - Added PD reusing of softmax and Batch Norm - Refactoring and removal of not needed routines of mkl-dnn ops test=develop - Fix to reusing conv test=develop - Lint fixes test=develop - Further lint fixes test=develop - Lint fixes test=develop - lint fixes test=develop - Lint workaround test=develop * - Fix after review on including boost as third party header test=develop * - Fix after review. Name change to something more descriptive test=develop
-
- 28 3月, 2019 1 次提交
-
-
由 Jacek Czaja 提交于
* Revert "[MKL-DNN] Fix to crash of Transformer when mkldnn is to be used (#16233)" This reverts commit 13816dd4. Apart from enabling transformer for MKL-DNN * Revert "- MKL-DNN pooling updated to set_prim_desc" This reverts commit c63f6b20. Conflicts: paddle/fluid/operators/mkldnn/concat_mkldnn_op.cc * Revert "[MKL-DNN] MKL-DNN specific Tensor modification (#15429)" test=develop This reverts commit dec9cf53. * - concat compilation fix - lint test=develop - Lint fixes test=develop - Lint fixes test=develop - Fix Transpose MKLDNN op test=develop
-
- 26 2月, 2019 1 次提交
-
-
由 Jacek Czaja 提交于
- MKLDNN ops revisited - disabled softmax modifications - disabled elementwise_add - reverted LRN modifications - reverted SUM primitive - Partial reviing of softmax - Enable softmax - Softmax changes - LRN is back - LRN partially disabled - LRN is back - LRN fix - compilation fixes - Sum fixed(hopefully) - Enabling (partially) elementwise_add - Fixes to elemenwise_add - Lint fixes quantize fix - compilation fix test=develop Disabling pooling - Disabled quantize op test=develop
-
- 29 1月, 2019 1 次提交
-
-
由 Krzysztof Binias 提交于
test=develop
-
- 27 11月, 2018 1 次提交
-
-
由 Jacek Czaja 提交于
test=develop - Added new header for MKLDNN reuse functionality - Extended conv2d_transpose GetExpectedKernelType for MKL-DNN supporrt - Buildable conv transpose mkldnn and conv mkldnn using conv template - Conv2d transpose roughlt implemented and buildable - Added modifications conv2d transpose MKLDNN unit tests - Fix to UT of conv2d transpose mkldnn op - Wrong type of MKLDNN primitive was chosen for conv2d transpose - HAcks for conv2d transpose - UT enalbed - Replaced copying loop with memcpy - Draft of passing lambda into AcquireMemory - Made reorder (IOHW->OIHW) to be called only once
-