From d6782b8856cd9e3f191bad2d36993df00eb8c8a5 Mon Sep 17 00:00:00 2001 From: cc <52520497+juncaipeng@users.noreply.github.com> Date: Fri, 21 Feb 2020 09:56:09 +0800 Subject: [PATCH] skip fusing quantized conv2d + relu6 for now (#2952) (#2957) skip fusing quantized conv2d + relu6 for now --- lite/core/mir/fusion/conv_activation_fuser.cc | 8 ++++++++ 1 file changed, 8 insertions(+) diff --git a/lite/core/mir/fusion/conv_activation_fuser.cc b/lite/core/mir/fusion/conv_activation_fuser.cc index 993fe4e944..37e259f3fa 100644 --- a/lite/core/mir/fusion/conv_activation_fuser.cc +++ b/lite/core/mir/fusion/conv_activation_fuser.cc @@ -53,6 +53,14 @@ void ConvActivationFuser::BuildPattern() { void ConvActivationFuser::InsertNewNode(SSAGraph* graph, const key2nodes_t& matched) { + // not fuse quantized conv2d + relu6 for now + auto conv2d_op_desc = matched.at("conv2d")->stmt()->op_info(); + bool is_conv2d_quantized = conv2d_op_desc->HasAttr("enable_int8") && + conv2d_op_desc->GetAttr("enable_int8"); + if (act_type_ == "relu6" && is_conv2d_quantized) { + return; + } + auto op_desc = GenOpDesc(matched); auto conv_op = LiteOpRegistry::Global().Create(conv_type_); auto conv_old = matched.at("conv2d")->stmt()->op(); -- GitLab