提交 9222d1a6 编写于 作者: S slf12

fix details

上级 3e618d7f
...@@ -170,7 +170,7 @@ def _quant_embedding_abs_max(graph, scope, place, config): ...@@ -170,7 +170,7 @@ def _quant_embedding_abs_max(graph, scope, place, config):
(1 << (bit_length - 1)) - 1)) (1 << (bit_length - 1)) - 1))
return scale, quanted_tensor.astype(np.int8) return scale, quanted_tensor.astype(np.int8)
def _insert_dequant_abx_max_op(graph, scope, var_node, scale_node, config): def _insert_dequant_abs_max_op(graph, scope, var_node, scale_node, config):
""" """
Insert dequantize_abs_max op in graph Insert dequantize_abs_max op in graph
""" """
...@@ -233,7 +233,7 @@ def _quant_embedding_abs_max(graph, scope, place, config): ...@@ -233,7 +233,7 @@ def _quant_embedding_abs_max(graph, scope, place, config):
if op_node.name() == 'lookup_table': if op_node.name() == 'lookup_table':
graph.update_input_link(embedding_node, quant_tensor_var, op_node) graph.update_input_link(embedding_node, quant_tensor_var, op_node)
var_node = op_node.outputs[0] var_node = op_node.outputs[0]
_insert_dequant_abx_max_op(graph, scope, var_node, scale_var, _insert_dequant_abs_max_op(graph, scope, var_node, scale_var,
config) config)
# free float embedding params memory # free float embedding params memory
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册