From 9f6049287e0cc3a31134ffccb623cfbfe59824c5 Mon Sep 17 00:00:00 2001 From: cc <52520497+juncaipeng@users.noreply.github.com> Date: Wed, 28 Jul 2021 15:45:34 +0800 Subject: [PATCH] quantize_transpiler_v2 supports quantize fp16 tensor (#34398) --- .../contrib/slim/quantization/quantize_transpiler_v2.py | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/python/paddle/fluid/contrib/slim/quantization/quantize_transpiler_v2.py b/python/paddle/fluid/contrib/slim/quantization/quantize_transpiler_v2.py index 753d68f7970..32768fff089 100644 --- a/python/paddle/fluid/contrib/slim/quantization/quantize_transpiler_v2.py +++ b/python/paddle/fluid/contrib/slim/quantization/quantize_transpiler_v2.py @@ -185,7 +185,10 @@ class QuantizeTranspilerV2(object): new_in_name = var_rename_map[block_id][in_name] else: in_var = block.var(in_name) - if in_var.dtype != core.VarDesc.VarType.FP32: + target_dtype = [ + core.VarDesc.VarType.FP32, core.VarDesc.VarType.FP16 + ] + if in_var.dtype not in target_dtype: continue quant_bits = self._weight_bits if in_var.persistable \ -- GitLab