From f450797421ba21ae984f21b53418d362d74ef7e3 Mon Sep 17 00:00:00 2001 From: WangXi Date: Thu, 28 Apr 2022 10:32:19 +0800 Subject: [PATCH] fix fused_multi_transformer compile failed in cuda arch < sm53 (#42315) --- paddle/fluid/operators/fused/fused_multi_transformer_op.cu | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/paddle/fluid/operators/fused/fused_multi_transformer_op.cu b/paddle/fluid/operators/fused/fused_multi_transformer_op.cu index f4a5319a68..e38ac9a0ad 100644 --- a/paddle/fluid/operators/fused/fused_multi_transformer_op.cu +++ b/paddle/fluid/operators/fused/fused_multi_transformer_op.cu @@ -534,6 +534,8 @@ template __global__ void masked_multihead_attention_kernel( Masked_multihead_attention_params params) { +#if CUDA_ARCH_FP16_SUPPORTED(__CUDA_ARCH__) + static_assert(Dh % THREADS_PER_KEY == 0, ""); static_assert(Dh % THREADS_PER_VALUE == 0, ""); @@ -821,6 +823,9 @@ __global__ void masked_multihead_attention_kernel( printf("\n"); } #endif +#else + assert(false); +#endif } template -- GitLab