未验证 提交 0c52e8a8 编写于 作者: R RichardWooSJTU 提交者: GitHub

skip this ut when cuda < 11.2 && cuda_arch < 8 (#49313)

上级 69e51c77
...@@ -15,8 +15,10 @@ ...@@ -15,8 +15,10 @@
import unittest import unittest
import numpy as np import numpy as np
from test_sparse_attention_op import get_cuda_version
import paddle import paddle
import paddle.fluid.core as core
import paddle.nn.functional as F import paddle.nn.functional as F
from paddle import _legacy_C_ops, tensor from paddle import _legacy_C_ops, tensor
from paddle.fluid.framework import default_main_program from paddle.fluid.framework import default_main_program
...@@ -128,6 +130,12 @@ def fused_multi_transformer_int8( ...@@ -128,6 +130,12 @@ def fused_multi_transformer_int8(
return final_out return final_out
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8Op(unittest.TestCase): class TestFusedMultiTransformerInt8Op(unittest.TestCase):
def setUp(self): def setUp(self):
self.config() self.config()
...@@ -781,6 +789,12 @@ class TestFusedMultiTransformerInt8Op(unittest.TestCase): ...@@ -781,6 +789,12 @@ class TestFusedMultiTransformerInt8Op(unittest.TestCase):
) )
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpFp16(TestFusedMultiTransformerInt8Op): class TestFusedMultiTransformerInt8OpFp16(TestFusedMultiTransformerInt8Op):
def config(self): def config(self):
super().config() super().config()
...@@ -788,6 +802,12 @@ class TestFusedMultiTransformerInt8OpFp16(TestFusedMultiTransformerInt8Op): ...@@ -788,6 +802,12 @@ class TestFusedMultiTransformerInt8OpFp16(TestFusedMultiTransformerInt8Op):
self.layers = 3 # odd layers self.layers = 3 # odd layers
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpCacheKV(TestFusedMultiTransformerInt8Op): class TestFusedMultiTransformerInt8OpCacheKV(TestFusedMultiTransformerInt8Op):
def config(self): def config(self):
super().config() super().config()
...@@ -798,6 +818,12 @@ class TestFusedMultiTransformerInt8OpCacheKV(TestFusedMultiTransformerInt8Op): ...@@ -798,6 +818,12 @@ class TestFusedMultiTransformerInt8OpCacheKV(TestFusedMultiTransformerInt8Op):
self.layers = 3 # odd layers self.layers = 3 # odd layers
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpCacheKVFp16( class TestFusedMultiTransformerInt8OpCacheKVFp16(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -809,6 +835,12 @@ class TestFusedMultiTransformerInt8OpCacheKVFp16( ...@@ -809,6 +835,12 @@ class TestFusedMultiTransformerInt8OpCacheKVFp16(
self.x_type = np.float16 self.x_type = np.float16
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpGenCacheKV( class TestFusedMultiTransformerInt8OpGenCacheKV(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -818,6 +850,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKV( ...@@ -818,6 +850,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKV(
self.gen_cache_kv = True self.gen_cache_kv = True
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpGenCacheKVFp16( class TestFusedMultiTransformerInt8OpGenCacheKVFp16(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -829,6 +867,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKVFp16( ...@@ -829,6 +867,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKVFp16(
self.layers = 3 # odd layers self.layers = 3 # odd layers
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpPostLayerNormFp16( class TestFusedMultiTransformerInt8OpPostLayerNormFp16(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -839,6 +883,12 @@ class TestFusedMultiTransformerInt8OpPostLayerNormFp16( ...@@ -839,6 +883,12 @@ class TestFusedMultiTransformerInt8OpPostLayerNormFp16(
self.pre_layer_norm = False self.pre_layer_norm = False
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpCacheKVPostLayerNorm( class TestFusedMultiTransformerInt8OpCacheKVPostLayerNorm(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -851,6 +901,12 @@ class TestFusedMultiTransformerInt8OpCacheKVPostLayerNorm( ...@@ -851,6 +901,12 @@ class TestFusedMultiTransformerInt8OpCacheKVPostLayerNorm(
self.pre_layer_norm = False self.pre_layer_norm = False
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpCacheKVPostLayerNormFp16( class TestFusedMultiTransformerInt8OpCacheKVPostLayerNormFp16(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -863,6 +919,12 @@ class TestFusedMultiTransformerInt8OpCacheKVPostLayerNormFp16( ...@@ -863,6 +919,12 @@ class TestFusedMultiTransformerInt8OpCacheKVPostLayerNormFp16(
self.pre_layer_norm = False self.pre_layer_norm = False
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNorm( class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNorm(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
...@@ -873,6 +935,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNorm( ...@@ -873,6 +935,12 @@ class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNorm(
self.pre_layer_norm = False self.pre_layer_norm = False
@unittest.skipIf(
not core.is_compiled_with_cuda()
or get_cuda_version() < 11020
or paddle.device.cuda.get_device_capability()[0] < 8,
"FusedMultiTransformerInt8 requires CUDA >= 11.2 and CUDA_ARCH >= 8",
)
class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNormFp16( class TestFusedMultiTransformerInt8OpGenCacheKVPostLayerNormFp16(
TestFusedMultiTransformerInt8Op TestFusedMultiTransformerInt8Op
): ):
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册