未验证 提交 04d2b4eb 编写于 作者: J juncaipeng 提交者: GitHub

set the bias of int8 conv as float, test=develop (#2553)

* set the bias of int8 conv as float, test=develop
上级 919282cf
...@@ -219,7 +219,7 @@ REGISTER_LITE_KERNEL(depthwise_conv2d, kARM, kFloat, kNCHW, ConvFp32, def) ...@@ -219,7 +219,7 @@ REGISTER_LITE_KERNEL(depthwise_conv2d, kARM, kFloat, kNCHW, ConvFp32, def)
REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out) REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out)
.BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) .BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt32))}) .BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kFloat))})
.BindInput("Filter", .BindInput("Filter",
{LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindOutput("Output", .BindOutput("Output",
...@@ -228,7 +228,7 @@ REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out) ...@@ -228,7 +228,7 @@ REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out)
REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out) REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out)
.BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) .BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt32))}) .BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kFloat))})
.BindInput("Filter", .BindInput("Filter",
{LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindOutput("Output", .BindOutput("Output",
...@@ -238,7 +238,7 @@ REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out) ...@@ -238,7 +238,7 @@ REGISTER_LITE_KERNEL(conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out)
REGISTER_LITE_KERNEL( REGISTER_LITE_KERNEL(
depthwise_conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out) depthwise_conv2d, kARM, kInt8, kNCHW, ConvInt8_Int8, int8_out)
.BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) .BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt32))}) .BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kFloat))})
.BindInput("Filter", .BindInput("Filter",
{LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindOutput("Output", .BindOutput("Output",
...@@ -248,7 +248,7 @@ REGISTER_LITE_KERNEL( ...@@ -248,7 +248,7 @@ REGISTER_LITE_KERNEL(
REGISTER_LITE_KERNEL( REGISTER_LITE_KERNEL(
depthwise_conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out) depthwise_conv2d, kARM, kInt8, kNCHW, ConvInt8_Fp32, fp32_out)
.BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) .BindInput("Input", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt32))}) .BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kFloat))})
.BindInput("Filter", .BindInput("Filter",
{LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))}) {LiteType::GetTensorTy(TARGET(kARM), PRECISION(kInt8))})
.BindOutput("Output", .BindOutput("Output",
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册