未验证 提交 cb68b58a 编写于 作者: N niuliling123 提交者: GitHub

Add fused_attention, fused_feedforward, fused_gemm_epilogue to amp white_list (#55373)

上级 bfb861f5
...@@ -23,6 +23,9 @@ FP16_WHITE_LIST = { ...@@ -23,6 +23,9 @@ FP16_WHITE_LIST = {
'mul', 'mul',
'fake_quantize_dequantize_abs_max', 'fake_quantize_dequantize_abs_max',
'fake_quantize_dequantize_moving_average_abs_max', 'fake_quantize_dequantize_moving_average_abs_max',
'fused_gemm_epilogue',
'fused_attention',
'fused_feedforward',
} }
# The set of ops that support fp16 calculation and are considered numerically- # The set of ops that support fp16 calculation and are considered numerically-
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册