- 28 3月, 2020 1 次提交
-
-
由 Wojciech Uss 提交于
-
- 12 2月, 2020 1 次提交
-
-
由 Wojciech Uss 提交于
* a test for Ernie QAT INT8 accuracy check test=develop * Remove NLP comparison test to split PRs test=develop * Fix typo and tabs, delete commented lines test=develop * re-combine the 2 PRs, test=develop Co-authored-by: NMichał Gallus <sand3r@interia.eu> Co-authored-by: Nbingyanghuang <33643817+bingyanghuang@users.noreply.github.com>
-
- 28 11月, 2019 1 次提交
-
-
由 lidanqing 提交于
* add ut for comparing FP32 and QAT INT8 * add save qat transformed model python script test=develop * updated * added missing file * add "with_label" test=develop * performance benchmark as unit test test=develop * change names of unnecessary thing * Change CMakeList.txt for model downloading and UT test=develop * change names of functions and params for more readable code test=develop * Change PADDLE_ENFORCE messages test=develop * fix indent problems test=develop * indent problems test=develop
-
- 14 11月, 2019 1 次提交
-
-
由 joanna.wozna.intel 提交于
test=develop
-
- 10 10月, 2019 1 次提交
-
-
由 Michał Gallus 提交于
* Leave fake quantization around mul * Replace Fake with Real Quantized Mul * Gather all scales from fake_quantize_ops * Enable uint8 in conv_relu tensors * Disable int8 mul and restore fake mul * Fix buf for running QAT on VGG16 and 19
-
- 28 9月, 2019 1 次提交
-
-
由 bingyanghuang 提交于
* Follow Wangzhen's comment in PR 18970, test=develop * Review comments, test=develop * Leave fake quantization around mul test=develop * Replace Fake with Real Quantized Mul test=develop * Fix bug in quantize placement pass Nodes in the graph now have checked type instead of node name when they are to be marked for quantization test=develop
-
- 25 9月, 2019 1 次提交
-
-
由 Wojciech Uss 提交于
* Add support for new QAT models test=develop Co-Authored-By: NMichał Gallus <michal.gallus@intel.com> Co-Authored-By: NWojciech Uss <wojciech.uss@intel.com> * fixed fps results test=develop * fix top5 accuracy drop problem * updated for new QAT models * skip quantizing average pooling - dirty but working * add missing pass * added missing conv+brelu fuse pass * removed a call to non-existent pass test=develop * renamed pass test=develop * Adjust finding pooling scale to newest QAT models * Remove unnecessary code from quantization_mkldnn_pass * Copy Pooling input scale to output scale in QAT * Refactor & remove unused code in QAT * Incorporate fp32 FC into QAT test=develop * Enable graph drawing with debug flag test=develop * Add tests for QATv2 * Fix paths for QATv2 models test=develop * Add option to save transformed int8 qat model test=develop * Remove redundant lines from qat mkldnn pass test=develop * Delegate disablement of avg pooling to qat test=develop * fix CI bug, test=develop * Follow Wangzhen's Review, test=develop * Update API.spec test=develop * Name False in (is_unsigned, TensorScale) tuple test=develop
-
- 09 7月, 2019 1 次提交
-
-
由 bingyanghuang 提交于
-
- 10 6月, 2019 1 次提交
-
-
由 bingyanghuang 提交于
-