未验证 提交 3c8722a4 编写于 作者: G Guanghua Yu 提交者: GitHub

cherry-pick #1477 #1524 (#1533)

上级 46f2957b
......@@ -126,9 +126,10 @@ def load_onnx_model(model_path,
v0, v1, v2 = version.split('.')
version_sum = int(v0) * 100 + int(v1) * 10 + int(v2)
if version_sum < 139:
_logger.error(
_logger.warning(
"x2paddle>=1.3.9 is required, please use \"pip install x2paddle\"."
)
os.system('python -m pip install -U x2paddle')
except:
os.system('python -m pip install -U x2paddle')
# check onnx installation and version
......
......@@ -118,7 +118,7 @@ class PTQ(object):
return fuse_list
def save_quantized_model(self, model, path, input_spec=None):
def save_quantized_model(self, model, path, input_spec=None, **kwargs):
"""
Save the quantized inference model.
......@@ -131,7 +131,7 @@ class PTQ(object):
InputSpec or example Tensor. If None, all input variables of
the original Layer's forward method would be the inputs of
the saved model. Default: None.
kwargs (dict, optional): Other save configuration options for compatibility.
Returns:
None
"""
......@@ -143,7 +143,7 @@ class PTQ(object):
model.eval()
self.ptq.save_quantized_model(
model=model, path=path, input_spec=input_spec)
model=model, path=path, input_spec=input_spec, **kwargs)
if training:
model.train()
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册