fix(dnn/cuda): disallow implicit dtype conversion in cublaslt matmul algos
disable tensor op matmul kernels when input and output tensors are in f32 data type to avoid potential accuracy loss GitOrigin-RevId: 36859cba5a526a7778e12c03ac32815144fe0505
Showing
想要评论请 注册 或 登录