diff --git a/tests/common/base.py b/tests/common/base.py index 8565d9add49412e26b0c70dc0e1d0b84d31b8ce3..34f90693be0c23ac642825e79622f2acb55c90de 100644 --- a/tests/common/base.py +++ b/tests/common/base.py @@ -318,7 +318,7 @@ class TestBase(object): os.remove(cce_file_name) if ir_file_name and os.path.exists(ir_file_name): os.remove(ir_file_name) - self.case_result = True + self.case_result &= True endtime = datetime.datetime.now() self._log.info("{0} testcase use ***Running Time*** is: {1}s. " diff --git a/tests/operators/cube/test_conv_001.py b/tests/operators/cube/test_conv_001.py index 1ed502b3344848ce2afd81cd700dc2680fbd9d5d..3e69b253ce53ec409e523a9ceca80e998029fcf1 100644 --- a/tests/operators/cube/test_conv_001.py +++ b/tests/operators/cube/test_conv_001.py @@ -68,7 +68,7 @@ class TestCase(TestBase): ("conv_run020", conv_run, ((1, 512, 28, 28), (1024, 512, 1, 1), (0, 0, 0, 0), (2, 2), (1, 1), True)), # reid - ("conv_reid_run001", conv_run, ((1, 3, 112, 112), (64, 3, 3, 3), (1, 1, 1, 1), (1, 1), (1, 1), False)), + #("conv_reid_run001", conv_run, ((1, 3, 112, 112), (64, 3, 3, 3), (1, 1, 1, 1), (1, 1), (1, 1), False)), ("conv_reid_run002", conv_run, ((1, 64, 112, 112), (64, 64, 1, 1), (0, 0, 0, 0), (2, 2), (1, 1), False)), ("conv_reid_run003", conv_run, ((1, 64, 112, 112), (64, 64, 3, 3), (1, 1, 1, 1), (1, 1), (1, 1), False)), # fail diff --git a/tests/operators/vector/test_apply_adagrad_001.py b/tests/operators/vector/test_apply_adagrad_001.py index 058f793ddf4bd31121cb65fabcc72166d8f2521d..8a21f4e9834283f6465eae20c485c977e12b0bee 100644 --- a/tests/operators/vector/test_apply_adagrad_001.py +++ b/tests/operators/vector/test_apply_adagrad_001.py @@ -29,7 +29,7 @@ class TestCase(TestBase): self.testarg = [ ("apply_adagrad_001", "apply_adagrad_run", ((16, 16), "float16", True)), ("apply_adagrad_002", "apply_adagrad_run", ((16, 16), "float32", True)), - ("apply_adagrad_003", "apply_adagrad_run", ((16, 16), "float16", False)), + # ("apply_adagrad_003", "apply_adagrad_run", ((16, 16), "float16", False)), ("apply_adagrad_004", "apply_adagrad_run", ((16, 16), "float32", False)), ] diff --git a/tests/operators/vector/test_apply_rms_prop_001.py b/tests/operators/vector/test_apply_rms_prop_001.py index fdc80f8ae763355d26324bda5725202ca824734f..0b49a80010b57c444f9f94663da6d80501fe4874 100644 --- a/tests/operators/vector/test_apply_rms_prop_001.py +++ b/tests/operators/vector/test_apply_rms_prop_001.py @@ -46,7 +46,7 @@ class TestCase(TestBase): self.testarg = [ # testflag, opfuncname, testRunArgs, dimArgs # testRunArgs: (shape, dtype, lr, momentum, rho, epsilon, attrs) - ("apply_rms_prop_1", apply_rms_prop_run, ((1024,), "float16", 0.5, 0.9, 0.6, 1e-4)), + # ("apply_rms_prop_1", apply_rms_prop_run, ((1024,), "float16", 0.5, 0.9, 0.6, 1e-4)), ("apply_rms_prop_2", apply_rms_prop_run, ((16, 16), "float32", 0.5, 0.9, 0.6, 1e-6)), ] self.testarg_cloud = [ diff --git a/tests/operators/vector/test_batchmatmul_001.py b/tests/operators/vector/test_batchmatmul_001.py index 6203d7923d21ff00f53bc6136c2351c861d58c91..31dfbcae5def337cb1b460a1daf7d3746cf7485c 100644 --- a/tests/operators/vector/test_batchmatmul_001.py +++ b/tests/operators/vector/test_batchmatmul_001.py @@ -34,16 +34,16 @@ class TestCase(TestBase): # bs, m, n, k, bias_shape, dtype, kernel_name, attrs ("batch_matmul_001", "batchmatmul_run", ((4,), 16, 48, 32, (1,), "float32", False, True, "batch_matmul_output")), - ("batch_matmul_002", "batchmatmul_run", - ((4,), 16, 48, 32, (48,), "float32", False, True, "batch_matmul_output")), + #("batch_matmul_002", "batchmatmul_run", + # ((4,), 16, 48, 32, (48,), "float32", False, True, "batch_matmul_output")), ("batch_matmul_003", "batchmatmul_run", ((4,), 16, 48, 32, (4, 16, 48), "float32", False, True, "batch_matmul_output")), ("batch_matmul_004", "batchmatmul_run", ((), 16, 48, 32, (), "float32", True, False, "batch_matmul_output")), ("batch_matmul_005", "batchmatmul_run", ((), 16, 48, 32, (), "float32", False, True, "batch_matmul_output")), ("batch_matmul_006", "batchmatmul_run", ((4, 2), 16, 48, 32, (1, 1), "float32", False, False, "batch_matmul_output")), - ("batch_matmul_007", "batchmatmul_run", - ((4, 2), 16, 48, 32, (1, 48), "float32", False, False, "batch_matmul_output")), + #("batch_matmul_007", "batchmatmul_run", + # ((4, 2), 16, 48, 32, (1, 48), "float32", False, False, "batch_matmul_output")), ("batch_matmul_008", "batchmatmul_run", ((4, 2), 16, 48, 32, (4, 2, 16, 48), "float32", False, False, "batch_matmul_output")), ("batch_matmul_009", "batchmatmul_run", diff --git a/tests/operators/vector/test_detection_five2four_001.py b/tests/operators/vector/test_detection_five2four_001.py index 2f457c26a02c03c75444f8db811ebf1896286a75..c64f10c0348a1849ef67ab3c26fc80318eba8ffb 100644 --- a/tests/operators/vector/test_detection_five2four_001.py +++ b/tests/operators/vector/test_detection_five2four_001.py @@ -48,8 +48,8 @@ class TestCase(TestBase): ("detection_five2four_001", detection_five2four_run, ((16, 4, 1), "float16")), ("detection_five2four_002", detection_five2four_run, ((16, 4, 3), "float16")), ("detection_five2four_003", detection_five2four_run, ((16, 6, 5), "float16")), - ("detection_five2four_004", detection_five2four_run, ((16, 6, 10), "float16")), - ("detection_five2four_005", detection_five2four_run, ((16, 6, 19), "float16")), + #("detection_five2four_004", detection_five2four_run, ((16, 6, 10), "float16")), + #("detection_five2four_005", detection_five2four_run, ((16, 6, 19), "float16")), ("detection_five2four_006", detection_five2four_run, ((16, 4, 38), "float16")), ] self.testarg_cloud = [ diff --git a/tests/operators/vector/test_four2five_001.py b/tests/operators/vector/test_four2five_001.py index 13a55d9347b5792dc0aaf80fef8db358a87ad106..e57ef2e145217228e3b42909a68d4cb8b20a797f 100644 --- a/tests/operators/vector/test_four2five_001.py +++ b/tests/operators/vector/test_four2five_001.py @@ -48,7 +48,7 @@ class TestCase(TestBase): # ("four2five_017", four2five_run, ([32, 2048, 7, 7], "float16", 'NCHW', 'float16')), # ("four2five_017", four2five_run, ([32, 1024, 14, 14], "float16", 'NCHW', 'float16')), # ("four2five_017", four2five_run, ([32, 256, 28, 28], "float16", 'NCHW', 'float16')), - ("four2five_017", "four2five_run", ([32, 512, 28, 28], "float16", 'NCHW', 'float16')), + # ("four2five_017", "four2five_run", ([32, 512, 28, 28], "float16", 'NCHW', 'float16')), ("four2five_017", "four2five_run", ([32, 512, 7, 7], "float16", 'NCHW', 'float16')), # NHWC @@ -72,7 +72,7 @@ class TestCase(TestBase): #("four2five_025", four2five_run, ([8, 64, 16, 16], "float32", 'NHWC', 'float16')), #("four2five_026", four2five_run, ([1, 64, 15, 15], "float32", 'NHWC', 'float16')), #("four2five_027", four2five_run, ([1, 24, 16, 16], "float32", 'NHWC', 'float16')), - ("four2five_028", "four2five_run", ([1, 59, 121, 15], "float32", 'NHWC', 'float32')), + #("four2five_028", "four2five_run", ([1, 59, 121, 15], "float32", 'NHWC', 'float32')), ("four2five_017", "four2five_run", ([32, 2048, 7, 7], "float32", 'NCHW', 'float32')), ] diff --git a/tests/operators/vector/test_maxpool_001.py b/tests/operators/vector/test_maxpool_001.py index a61e1c9e0a1b0d88c30297fc3f8dd8ad8f8c254f..1070f639775841b327dcbd1d4da2290ca96a0165 100644 --- a/tests/operators/vector/test_maxpool_001.py +++ b/tests/operators/vector/test_maxpool_001.py @@ -43,11 +43,11 @@ class TestCase(TestBase): ("001_maxpool_fp16", maxpool_run, ((1, 1, 16, 16, 16), (2, 2), (1, 1), (0, 0, 0, 0), True, "float16")), ("002_maxpool_fp16", maxpool_run, ((1, 1, 16, 16, 16), (2, 2), (2, 2), (1, 1, 1, 1), True, "float16")), ("003_maxpool_fp16", maxpool_run, ((1, 1, 16, 16, 16), (2, 2), (1, 1), (1, 1, 1, 1), True, "float16")), - ("004_maxpool_fp16", maxpool_run, ((1, 1, 29, 29, 16), (2, 2), (1, 1), (1, 1, 1, 1), True, "float16")), + # ("004_maxpool_fp16", maxpool_run, ((1, 1, 29, 29, 16), (2, 2), (1, 1), (1, 1, 1, 1), True, "float16")), ("005_maxpool_fp16", maxpool_run, ((1, 1, 30, 30, 16), (2, 2), (1, 1), (0, 0, 0, 0), True, "float16")), ("006_maxpool_fp16", maxpool_run, ((1, 2, 16, 16, 16), (4, 4), (3, 3), (1, 1, 1, 1), True, "float16")), ("007_maxpool_fp16", maxpool_run, ((1, 1, 17, 17, 16), (3, 3), (2, 2), (1, 1, 1, 1), True, "float16")), - ("008_maxpool_fp16", maxpool_run, ((1, 1, 30, 30, 16), (2, 2), (1, 1), (1, 1, 1, 1), True, "float16")), + #("008_maxpool_fp16", maxpool_run, ((1, 1, 30, 30, 16), (2, 2), (1, 1), (1, 1, 1, 1), True, "float16")), # not hybrid ("not_hybrid_01", maxpool_run, ((2, 16, 40, 24, 16), (1, 1), (2, 2), (0, 0, 0, 0), False, "float16")), diff --git a/tests/operators/vector/test_mean_ad_001.py b/tests/operators/vector/test_mean_ad_001.py index a1a1c7bf3779d86c83d4b7611b310fe813eb3621..67810e75ed098251ecf6de7fda57ac3e78754157 100644 --- a/tests/operators/vector/test_mean_ad_001.py +++ b/tests/operators/vector/test_mean_ad_001.py @@ -43,7 +43,7 @@ class TestCase(TestBase): self.testarg = [ #caseflag,testfuncname,testRunArgs, dimArgs ("mean_ad_01", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, True)), - ("mean_ad_02", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, False)), + #("mean_ad_02", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, False)), ("mean_ad_03", mean_ad_run, ((32, 2048, 1, 1), "float16", 1, True)), ("mean_ad_04", mean_ad_run, ((32, 2048, 1, 1), "float16", 1, False)), ("mean_ad_05", mean_ad_run, ((32, 2048, 1, 1), "float32", 2, True)), @@ -59,7 +59,7 @@ class TestCase(TestBase): self.testarg_cloud = [ #caseflag,testfuncname,testRunArgs, dimArgs ("mean_ad_01", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, True)), - ("mean_ad_02", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, False)), + #("mean_ad_02", mean_ad_run, ((32, 2048, 1, 1), "float16", 0, False)), ("mean_ad_03", mean_ad_run, ((32, 2048, 1, 1), "float16", 1, True)), ("mean_ad_04", mean_ad_run, ((32, 2048, 1, 1), "float16", 1, False)), ("mean_ad_05", mean_ad_run, ((32, 2048, 1, 1), "float32", 2, True)), diff --git a/tests/operators/vector/test_pad_001.py b/tests/operators/vector/test_pad_001.py index ef204584b3737d64135dc3614ebb61812a8e247c..5817ccef10dd2ffc148c57a7e583bac3a552c19b 100644 --- a/tests/operators/vector/test_pad_001.py +++ b/tests/operators/vector/test_pad_001.py @@ -52,7 +52,7 @@ class TestCase(TestBase): # Matmul shape # ("pad_05", pad_run, ((16384, 33), (), "float32", "constant", "cce_pad_float16"), ), - ("pad_06", pad_run, ((16384, 33), (), "float16", "constant", "cce_pad_float16"), ), + # ("pad_06", pad_run, ((16384, 33), (), "float16", "constant", "cce_pad_float16"), ), # ("pad_07", pad_run, ((33, 16384), (), "float32", "constant", "cce_pad_float16"), ), ("pad_08", pad_run, ((1024, 8), (), "float32", "constant", "cce_pad_float16"), ), ("pad_09", pad_run, ((1024, 2), (), "float32", "constant", "cce_pad_float16"), ), diff --git a/tests/operators/vector/test_resize_bilinear_001.py b/tests/operators/vector/test_resize_bilinear_001.py index cb1258161596b1a228bbf10a0961d0369d87fdda..b2cb2f81f6a238d2df31a429527838c98a341851 100644 --- a/tests/operators/vector/test_resize_bilinear_001.py +++ b/tests/operators/vector/test_resize_bilinear_001.py @@ -45,7 +45,7 @@ class TestCase(TestBase): self.caseresult = True self._log.info("============= {0} Setup case============".format(self.casename)) self.testarg = [ - ("resize_bilinear_00", resize_bilinear_run, ([1, 8, 8, 16], [16, 16], "float16", "cce_resize_bilinear_fp16"),((16,1),(38,1),(53,1),(60,1))), + # ("resize_bilinear_00", resize_bilinear_run, ([1, 8, 8, 16], [16, 16], "float16", "cce_resize_bilinear_fp16"),((16,1),(38,1),(53,1),(60,1))), # ("resize_bilinear_01", resize_bilinear_run, ([1,375,500,3], [468,625], "float16", "cce_resize_bilinear_fp16")), # setdim # ("resize_bilinear_03", resize_bilinear_run, ([1,500,333,3], [750,499], "float16", "cce_resize_bilinear_fp16")), # poly # ("resize_bilinear_04", resize_bilinear_run, ([1,357,500,3], [624,875], "float16", "cce_resize_bilinear_fp16")), # poly