INT8 inference accuracy lost after enabling quantization of elementwise_add in a new Ernie Quant model
Created by: wojtuss
Recently @lidanqing-intel has generated a new Quant Ernie model. It contains quantization scales for elementwise_add
operators to be quantized. However, after enabling quantization of elementwise_add
ops, the accuracy is lost.
We are investigating the issue.