diff --git a/python/paddle/fluid/tests/unittests/test_tril_indices_op.py b/python/paddle/fluid/tests/unittests/test_tril_indices_op.py index 98236ee235ddf4e2eb36fb39f2cd9d9c5fb716ad..6c473f2a8a817fec3a8555a946f26764d0055a03 100644 --- a/python/paddle/fluid/tests/unittests/test_tril_indices_op.py +++ b/python/paddle/fluid/tests/unittests/test_tril_indices_op.py @@ -19,7 +19,6 @@ from op_test import OpTest import paddle import paddle.fluid as fluid -from paddle.fluid.framework import _test_eager_guard class TestTrilIndicesOp(OpTest): @@ -89,10 +88,6 @@ class TestTrilIndicesAPICaseDygraph(unittest.TestCase): expected_result1 = np.tril_indices(4, 2, 4) self.assertEqual((out1.numpy() == expected_result1).all(), True) - def test_dygraph_eager(self): - with _test_eager_guard(): - self.test_dygraph() - class TestTrilIndicesAPICaseError(unittest.TestCase): def test_case_error(self): diff --git a/python/paddle/fluid/tests/unittests/test_triu_indices_op.py b/python/paddle/fluid/tests/unittests/test_triu_indices_op.py index a666d63b7d006737e4dbb88e0c87cc2d8cc678ba..a9e391d6adf2612f72dc2c492d8fb2ab8bb6c332 100644 --- a/python/paddle/fluid/tests/unittests/test_triu_indices_op.py +++ b/python/paddle/fluid/tests/unittests/test_triu_indices_op.py @@ -19,7 +19,6 @@ from op_test import OpTest import paddle import paddle.fluid as fluid -from paddle.fluid.framework import _test_eager_guard class TestTriuIndicesOp(OpTest): @@ -84,10 +83,6 @@ class TestTriuIndicesAPICaseDygraph(unittest.TestCase): expected_result = np.triu_indices(4, 2, 4) np.testing.assert_array_equal(out, expected_result) - def test_dygraph_eager(self): - with _test_eager_guard(): - self.test_dygraph() - class TestTriuIndicesAPICaseError(unittest.TestCase): def test_case_error(self): diff --git a/python/paddle/fluid/tests/unittests/test_trunc_op.py b/python/paddle/fluid/tests/unittests/test_trunc_op.py index 664cf00a21b8417d71c6659ba773c1b3a4bc421d..db45b36b563020f388a933ea369964deaed7f4ac 100644 --- a/python/paddle/fluid/tests/unittests/test_trunc_op.py +++ b/python/paddle/fluid/tests/unittests/test_trunc_op.py @@ -18,7 +18,6 @@ import numpy as np from op_test import OpTest import paddle -from paddle.fluid.framework import _test_eager_guard paddle.enable_static() @@ -85,20 +84,6 @@ class TestTruncAPI(unittest.TestCase): np.testing.assert_allclose(out.numpy(), out_ref, rtol=1e-08) paddle.enable_static() - def test_api_eager(self): - paddle.disable_static(self.place) - - with _test_eager_guard(): - x_tensor = paddle.to_tensor(self.x) - out = paddle.trunc(x_tensor) - out_ref = np.trunc(self.x) - np.testing.assert_allclose(out.numpy(), out_ref, rtol=1e-08) - paddle.enable_static() - - def test_api_eager_dygraph(self): - with _test_eager_guard(): - self.test_api_dygraph() - def test_errors(self): with paddle.static.program_guard(paddle.static.Program()): x = paddle.fluid.data('X', [20, 20], 'bool') diff --git a/python/paddle/fluid/tests/unittests/test_truncated_gaussian_random_op.py b/python/paddle/fluid/tests/unittests/test_truncated_gaussian_random_op.py index 941e56b808efd56d086fbb7ca2030453b57f2478..30a91cc23093f9fbaa45587728ee0035eabc9886 100644 --- a/python/paddle/fluid/tests/unittests/test_truncated_gaussian_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_truncated_gaussian_random_op.py @@ -20,7 +20,6 @@ import paddle import paddle.fluid as fluid import paddle.fluid.core as core from paddle.fluid.executor import Executor -from paddle.fluid.framework import _test_eager_guard class TestTrunctedGaussianRandomOp(unittest.TestCase): @@ -70,17 +69,16 @@ class TestTrunctedGaussianRandomOp(unittest.TestCase): # directly def gaussian_random_test_eager(self, place): with fluid.dygraph.guard(place): - with _test_eager_guard(): - out = paddle._C_ops.truncated_gaussian_random( - self.attrs["shape"], - self.attrs["mean"], - self.attrs["std"], - self.attrs["seed"], - core.VarDesc.VarType.FP32, - place, - ) - self.assertAlmostEqual(numpy.mean(out.numpy()), 0.0, delta=0.1) - self.assertAlmostEqual(numpy.var(out.numpy()), 0.773, delta=0.1) + out = paddle._C_ops.truncated_gaussian_random( + self.attrs["shape"], + self.attrs["mean"], + self.attrs["std"], + self.attrs["seed"], + core.VarDesc.VarType.FP32, + place, + ) + self.assertAlmostEqual(numpy.mean(out.numpy()), 0.0, delta=0.1) + self.assertAlmostEqual(numpy.var(out.numpy()), 0.773, delta=0.1) if __name__ == "__main__": diff --git a/python/paddle/fluid/tests/unittests/test_unbind_op.py b/python/paddle/fluid/tests/unittests/test_unbind_op.py index 6f49d66b1aca60bb78184c66e47c86761a64e979..6ec82a96bc16534572d59eb81c690839681be557 100644 --- a/python/paddle/fluid/tests/unittests/test_unbind_op.py +++ b/python/paddle/fluid/tests/unittests/test_unbind_op.py @@ -21,7 +21,6 @@ import paddle import paddle.fluid as fluid import paddle.tensor as tensor from paddle.fluid import Program, program_guard -from paddle.fluid.framework import _test_eager_guard class TestUnbind(unittest.TestCase): @@ -57,10 +56,6 @@ class TestUnbind(unittest.TestCase): out.backward() np.testing.assert_array_equal(x.grad.numpy(), np_grad) - def test_unbind_dygraph_final_state(self): - with _test_eager_guard(): - self.test_unbind_dygraph() - class TestLayersUnbind(unittest.TestCase): def test_layers_unbind(self): diff --git a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py index 86e932d25bf03807330c815189990f2e35ff0ea3..6f7ffc9437639850331f84c890b6e457cbad52f1 100644 --- a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py @@ -23,7 +23,6 @@ import paddle import paddle.fluid as fluid import paddle.fluid.core as core from paddle.fluid import Program, program_guard -from paddle.fluid.framework import _test_eager_guard from paddle.fluid.op import Operator from paddle.tensor import random @@ -171,7 +170,7 @@ class TestUniformRandomOp(OpTest): hist, prob = self.output_hist(np.array(outs[0])) np.testing.assert_allclose(hist, prob, rtol=0, atol=0.01) - def func_test_check_api(self): + def test_check_api(self): places = self._get_places() for place in places: with fluid.dygraph.base.guard(place=place): @@ -183,11 +182,6 @@ class TestUniformRandomOp(OpTest): self.attrs['seed'], ) - def test_check_api_eager(self): - with _test_eager_guard(): - self.func_test_check_api() - self.func_test_check_api() - class TestUniformRandomOpError(unittest.TestCase): def test_errors(self): diff --git a/python/paddle/fluid/tests/unittests/test_unique.py b/python/paddle/fluid/tests/unittests/test_unique.py index d454d96a446cbefe4aa8381b8992d914b56f381a..3ff4bae9b26db69af8f18297ea582b9fa35acb12 100644 --- a/python/paddle/fluid/tests/unittests/test_unique.py +++ b/python/paddle/fluid/tests/unittests/test_unique.py @@ -20,7 +20,6 @@ from op_test import OpTest import paddle import paddle.fluid as fluid import paddle.fluid.core as core -from paddle.fluid.framework import _test_eager_guard class TestUniqueOp(OpTest): @@ -262,12 +261,6 @@ class TestUniqueAPI(unittest.TestCase): self.assertTrue((counts.numpy() == np_counts).all(), True) paddle.enable_static() - def test_dygraph_api(self): - with _test_eager_guard(): - self.test_dygraph_api_out() - self.test_dygraph_api_attr() - self.test_dygraph_attr_dtype() - def test_static_graph(self): with paddle.static.program_guard( paddle.static.Program(), paddle.static.Program() diff --git a/python/paddle/fluid/tests/unittests/test_var_base.py b/python/paddle/fluid/tests/unittests/test_var_base.py index e7fbd7bf81440ac1ac4360cf11e0cfe9a4b44d68..f542f51bc2479dcb10c8c4f0eef83e1d402639de 100644 --- a/python/paddle/fluid/tests/unittests/test_var_base.py +++ b/python/paddle/fluid/tests/unittests/test_var_base.py @@ -21,7 +21,7 @@ import paddle import paddle.fluid as fluid import paddle.fluid.core as core import paddle.nn.functional as F -from paddle.fluid.framework import _in_legacy_dygraph, _test_eager_guard +from paddle.fluid.framework import _in_legacy_dygraph class TestVarBase(unittest.TestCase): @@ -30,7 +30,7 @@ class TestVarBase(unittest.TestCase): self.dtype = np.float32 self.array = np.random.uniform(0.1, 1, self.shape).astype(self.dtype) - def func_test_to_tensor(self): + def test_to_tensor(self): def check_with_place(place): with fluid.dygraph.guard(): paddle.set_default_dtype('float32') @@ -277,12 +277,7 @@ class TestVarBase(unittest.TestCase): check_with_place(core.NPUPlace(0)) check_with_place("npu:0") - def test_to_tensor(self): - with _test_eager_guard(): - self.func_test_to_tensor() - self.func_test_to_tensor() - - def func_test_to_tensor_not_change_input_stop_gradient(self): + def test_to_tensor_not_change_input_stop_gradient(self): with paddle.fluid.dygraph.guard(core.CPUPlace()): a = paddle.zeros([1024]) a.stop_gradient = False @@ -290,12 +285,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a.stop_gradient, False) self.assertEqual(b.stop_gradient, True) - def test_to_tensor_not_change_input_stop_gradient(self): - with _test_eager_guard(): - self.func_test_to_tensor_not_change_input_stop_gradient() - self.func_test_to_tensor_not_change_input_stop_gradient() - - def func_test_to_tensor_change_place(self): + def test_to_tensor_change_place(self): if core.is_compiled_with_cuda(): a_np = np.random.rand(1024, 1024) with paddle.fluid.dygraph.guard(core.CPUPlace()): @@ -313,12 +303,7 @@ class TestVarBase(unittest.TestCase): a = paddle.to_tensor(a, place=paddle.CUDAPinnedPlace()) self.assertEqual(a.place.__repr__(), "Place(gpu_pinned)") - def test_to_tensor_change_place(self): - with _test_eager_guard(): - self.func_test_to_tensor_change_place() - self.func_test_to_tensor_change_place() - - def func_test_to_tensor_with_lodtensor(self): + def test_to_tensor_with_lodtensor(self): if core.is_compiled_with_cuda(): a_np = np.random.rand(1024, 1024) with paddle.fluid.dygraph.guard(core.CPUPlace()): @@ -334,12 +319,7 @@ class TestVarBase(unittest.TestCase): np.testing.assert_array_equal(a_np, a.numpy()) self.assertTrue(a.place.__repr__(), "Place(cpu)") - def test_to_tensor_with_lodtensor(self): - with _test_eager_guard(): - self.func_test_to_tensor_with_lodtensor() - self.func_test_to_tensor_with_lodtensor() - - def func_test_to_variable(self): + def test_to_variable(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array, name="abc") np.testing.assert_array_equal(var.numpy(), self.array) @@ -358,12 +338,7 @@ class TestVarBase(unittest.TestCase): linear = paddle.nn.Linear(32, 64) var = linear._helper.to_variable("test", name="abc") - def test_to_variable(self): - with _test_eager_guard(): - self.func_test_to_variable() - self.func_test_to_variable() - - def func_test_list_to_variable(self): + def test_list_to_variable(self): with fluid.dygraph.guard(): array = [[[1, 2], [1, 2], [1.0, 2]], [[1, 2], [1, 2], [1, 2]]] var = fluid.dygraph.to_variable(array, dtype='int32') @@ -372,12 +347,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(var.dtype, core.VarDesc.VarType.INT32) self.assertEqual(var.type, core.VarDesc.VarType.LOD_TENSOR) - def test_list_to_variable(self): - with _test_eager_guard(): - self.func_test_list_to_variable() - self.func_test_list_to_variable() - - def func_test_tuple_to_variable(self): + def test_tuple_to_variable(self): with fluid.dygraph.guard(): array = (((1, 2), (1, 2), (1, 2)), ((1, 2), (1, 2), (1, 2))) var = fluid.dygraph.to_variable(array, dtype='float32') @@ -386,24 +356,14 @@ class TestVarBase(unittest.TestCase): self.assertEqual(var.dtype, core.VarDesc.VarType.FP32) self.assertEqual(var.type, core.VarDesc.VarType.LOD_TENSOR) - def test_tuple_to_variable(self): - with _test_eager_guard(): - self.func_test_tuple_to_variable() - self.func_test_tuple_to_variable() - - def func_test_tensor_to_variable(self): + def test_tensor_to_variable(self): with fluid.dygraph.guard(): t = fluid.Tensor() t.set(np.random.random((1024, 1024)), fluid.CPUPlace()) var = fluid.dygraph.to_variable(t) np.testing.assert_array_equal(t, var.numpy()) - def test_tensor_to_variable(self): - with _test_eager_guard(): - self.func_test_tensor_to_variable() - self.func_test_tensor_to_variable() - - def func_test_leaf_tensor(self): + def test_leaf_tensor(self): with fluid.dygraph.guard(): x = paddle.to_tensor(np.random.uniform(-1, 1, size=[10, 10])) self.assertTrue(x.is_leaf) @@ -429,12 +389,7 @@ class TestVarBase(unittest.TestCase): self.assertTrue(linear.bias.is_leaf) self.assertFalse(out.is_leaf) - def test_leaf_tensor(self): - with _test_eager_guard(): - self.func_test_leaf_tensor() - self.func_test_leaf_tensor() - - def func_test_detach(self): + def test_detach(self): with fluid.dygraph.guard(): x = paddle.to_tensor(1.0, dtype="float64", stop_gradient=False) detach_x = x.detach() @@ -470,12 +425,7 @@ class TestVarBase(unittest.TestCase): detach_x[:] = 5.0 y.backward() - def test_detach(self): - with _test_eager_guard(): - self.func_test_detach() - self.func_test_detach() - - def func_test_write_property(self): + def test_write_property(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) @@ -491,12 +441,7 @@ class TestVarBase(unittest.TestCase): var.stop_gradient = False self.assertEqual(var.stop_gradient, False) - def test_write_property(self): - with _test_eager_guard(): - self.func_test_write_property() - self.func_test_write_property() - - def func_test_deep_copy(self): + def test_deep_copy(self): with fluid.dygraph.guard(): if _in_legacy_dygraph(): empty_var = core.VarBase() @@ -579,13 +524,8 @@ class TestVarBase(unittest.TestCase): np.array(selected_rows.get_tensor()), ) - def test_deep_copy(self): - with _test_eager_guard(): - self.func_test_deep_copy() - self.func_test_deep_copy() - # test some patched methods - def func_test_set_value(self): + def test_set_value(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) tmp1 = np.random.uniform(0.1, 1, [2, 2, 3]).astype(self.dtype) @@ -595,22 +535,12 @@ class TestVarBase(unittest.TestCase): var.set_value(tmp2) np.testing.assert_array_equal(var.numpy(), tmp2) - def test_set_value(self): - with _test_eager_guard(): - self.func_test_set_value() - self.func_test_set_value() - - def func_test_to_string(self): + def test_to_string(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) self.assertTrue(isinstance(str(var), str)) - def test_to_string(self): - with _test_eager_guard(): - self.func_test_to_string() - self.func_test_to_string() - - def func_test_element_size(self): + def test_element_size(self): with fluid.dygraph.guard(): x = paddle.to_tensor(1, dtype='bool') self.assertEqual(x.element_size(), 1) @@ -645,12 +575,7 @@ class TestVarBase(unittest.TestCase): x = paddle.to_tensor(1, dtype='complex128') self.assertEqual(x.element_size(), 16) - def test_element_size(self): - with _test_eager_guard(): - self.func_test_element_size() - self.func_test_element_size() - - def func_test_backward(self): + def test_backward(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) var.stop_gradient = False @@ -659,12 +584,7 @@ class TestVarBase(unittest.TestCase): grad_var = var._grad_ivar() self.assertEqual(grad_var.shape, self.shape) - def test_backward(self): - with _test_eager_guard(): - self.func_test_backward() - self.func_test_backward() - - def func_test_gradient(self): + def test_gradient(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) var.stop_gradient = False @@ -673,23 +593,13 @@ class TestVarBase(unittest.TestCase): grad_var = var.gradient() self.assertEqual(grad_var.shape, self.array.shape) - def test_gradient(self): - with _test_eager_guard(): - self.func_test_gradient() - self.func_test_gradient() - - def func_test_block(self): + def test_block(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) self.assertEqual( var.block, fluid.default_main_program().global_block() ) - def test_block(self): - with _test_eager_guard(): - self.func_test_block() - self.func_test_block() - def _test_slice(self): w = fluid.dygraph.to_variable( np.random.random((784, 100, 100)).astype('float64') @@ -1079,7 +989,7 @@ class TestVarBase(unittest.TestCase): col = np.array([2, 1, 3]) np.testing.assert_array_equal(array[row, col], x[row, col].numpy()) - def func_test_slice(self): + def test_slice(self): with fluid.dygraph.guard(): self._test_slice() self._test_slice_for_tensor_attr() @@ -1105,24 +1015,14 @@ class TestVarBase(unittest.TestCase): mask = np.array([1, 0, 1, 0], dtype=bool) var[paddle.to_tensor([0, 1]), mask] - def test_slice(self): - with _test_eager_guard(): - self.func_test_slice() - self.func_test_slice() - - def func_test_var_base_to_np(self): + def test_var_base_to_np(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) np.testing.assert_array_equal( var.numpy(), fluid.framework._var_base_to_np(var) ) - def test_var_base_to_np(self): - with _test_eager_guard(): - self.func_test_var_base_to_np() - self.func_test_var_base_to_np() - - def func_test_var_base_as_np(self): + def test_var_base_as_np(self): with fluid.dygraph.guard(): var = fluid.dygraph.to_variable(self.array) np.testing.assert_array_equal(var.numpy(), np.array(var)) @@ -1130,12 +1030,7 @@ class TestVarBase(unittest.TestCase): var.numpy(), np.array(var, dtype=np.float32) ) - def test_var_base_as_np(self): - with _test_eager_guard(): - self.func_test_var_base_as_np() - self.func_test_var_base_as_np() - - def func_test_if(self): + def test_if(self): with fluid.dygraph.guard(): var1 = fluid.dygraph.to_variable(np.array([[[0]]])) var2 = fluid.dygraph.to_variable(np.array([[[1]]])) @@ -1154,12 +1049,7 @@ class TestVarBase(unittest.TestCase): assert not bool(var1), "bool(var1) is False" assert bool(var2), "bool(var2) is True" - def test_if(self): - with _test_eager_guard(): - self.func_test_if() - self.func_test_if() - - def func_test_to_static_var(self): + def test_to_static_var(self): with fluid.dygraph.guard(): # Convert VarBase into Variable or Parameter var_base = fluid.dygraph.to_variable(self.array, name="var_base_1") @@ -1184,11 +1074,6 @@ class TestVarBase(unittest.TestCase): static_param = weight._to_static_var() self._assert_to_static(weight, static_param, True) - def test_to_static_var(self): - with _test_eager_guard(): - self.func_test_to_static_var() - self.func_test_to_static_var() - def _assert_to_static(self, var_base, static_var, is_param=False): if is_param: self.assertTrue(isinstance(static_var, fluid.framework.Parameter)) @@ -1216,7 +1101,7 @@ class TestVarBase(unittest.TestCase): self.assertListEqual(list(var_base.shape), list(static_var.shape)) - def func_test_tensor_str(self): + def test_tensor_str(self): paddle.enable_static() paddle.disable_static(paddle.CPUPlace()) paddle.seed(10) @@ -1235,12 +1120,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str(self): - with _test_eager_guard(): - self.func_test_tensor_str() - self.func_test_tensor_str() - - def func_test_tensor_str2(self): + def test_tensor_str2(self): paddle.disable_static(paddle.CPUPlace()) a = paddle.to_tensor([[1.5111111, 1.0], [0, 0]]) a_str = str(a) @@ -1251,12 +1131,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str2(self): - with _test_eager_guard(): - self.func_test_tensor_str2() - self.func_test_tensor_str2() - - def func_test_tensor_str3(self): + def test_tensor_str3(self): paddle.disable_static(paddle.CPUPlace()) a = paddle.to_tensor([[-1.5111111, 1.0], [0, -0.5]]) a_str = str(a) @@ -1267,12 +1142,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str3(self): - with _test_eager_guard(): - self.func_test_tensor_str3() - self.func_test_tensor_str3() - - def func_test_tensor_str_scaler(self): + def test_tensor_str_scaler(self): paddle.disable_static(paddle.CPUPlace()) a = paddle.to_tensor(np.array(False)) a_str = str(a) @@ -1282,12 +1152,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str_scaler(self): - with _test_eager_guard(): - self.func_test_tensor_str_scaler() - self.func_test_tensor_str_scaler() - - def func_test_tensor_str_shape_with_zero(self): + def test_tensor_str_shape_with_zero(self): paddle.disable_static(paddle.CPUPlace()) x = paddle.ones((10, 10)) y = paddle.nonzero(x == 0) @@ -1298,12 +1163,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str_shape_with_zero(self): - with _test_eager_guard(): - self.func_test_tensor_str_shape_with_zero() - self.func_test_tensor_str_shape_with_zero() - - def func_test_tensor_str_linewidth(self): + def test_tensor_str_linewidth(self): paddle.disable_static(paddle.CPUPlace()) paddle.seed(2021) x = paddle.rand([128]) @@ -1331,12 +1191,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str_linewidth(self): - with _test_eager_guard(): - self.func_test_tensor_str_linewidth() - self.func_test_tensor_str_linewidth() - - def func_test_tensor_str_linewidth2(self): + def test_tensor_str_linewidth2(self): paddle.disable_static(paddle.CPUPlace()) paddle.seed(2021) x = paddle.rand([128]) @@ -1358,12 +1213,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str_linewidth2(self): - with _test_eager_guard(): - self.func_test_tensor_str_linewidth2() - self.func_test_tensor_str_linewidth2() - - def func_tensor_str_bf16(self): + def test_tensor_str_bf16(self): paddle.disable_static(paddle.CPUPlace()) a = paddle.to_tensor([[1.5, 1.0], [0, 0]]) a = paddle.cast(a, dtype=core.VarDesc.VarType.BF16) @@ -1376,12 +1226,7 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_tensor_str_bf16(self): - with _test_eager_guard(): - self.func_tensor_str_bf16() - self.func_tensor_str_bf16() - - def func_test_print_tensor_dtype(self): + def test_print_tensor_dtype(self): paddle.disable_static(paddle.CPUPlace()) a = paddle.rand([1]) a_str = str(a.dtype) @@ -1390,11 +1235,6 @@ class TestVarBase(unittest.TestCase): self.assertEqual(a_str, expected) - def test_print_tensor_dtype(self): - with _test_eager_guard(): - self.func_test_print_tensor_dtype() - self.func_test_print_tensor_dtype() - class TestVarBaseSetitem(unittest.TestCase): def func_setUp(self): @@ -1440,9 +1280,6 @@ class TestVarBaseSetitem(unittest.TestCase): self._test(self.tensor_value) def test_value_tensor(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_tensor() self.func_setUp() self.func_test_value_tensor() @@ -1450,9 +1287,6 @@ class TestVarBaseSetitem(unittest.TestCase): self._test(self.np_value) def test_value_numpy(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_numpy() self.func_setUp() self.func_test_value_numpy() @@ -1460,9 +1294,6 @@ class TestVarBaseSetitem(unittest.TestCase): self._test(10) def test_value_int(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_int() self.func_setUp() self.func_test_value_int() @@ -1481,9 +1312,6 @@ class TestVarBaseSetitemFp32(TestVarBaseSetitem): self._test(3.3) def test_value_float(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_float() self.func_setUp() self.func_test_value_float() @@ -1542,9 +1370,6 @@ class TestVarBaseSetitemBoolIndex(unittest.TestCase): self._test(self.tensor_value) def test_value_tensor(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_tensor() self.func_setUp() self.func_test_value_tensor() @@ -1553,9 +1378,6 @@ class TestVarBaseSetitemBoolIndex(unittest.TestCase): self._test(self.np_value) def test_value_numpy(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_numpy() self.func_setUp() self.func_test_value_numpy() @@ -1564,9 +1386,6 @@ class TestVarBaseSetitemBoolIndex(unittest.TestCase): self._test(10) def test_value_int(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_value_int() self.func_setUp() self.func_test_value_int() @@ -1597,7 +1416,7 @@ class TestVarBaseSetitemBoolScalarIndex(unittest.TestCase): class TestVarBaseInplaceVersion(unittest.TestCase): - def func_test_setitem(self): + def test_setitem(self): paddle.disable_static() var = paddle.ones(shape=[4, 2, 3], dtype="float32") @@ -1609,12 +1428,7 @@ class TestVarBaseInplaceVersion(unittest.TestCase): var[1:2] = 1 self.assertEqual(var.inplace_version, 2) - def test_setitem(self): - with _test_eager_guard(): - self.func_test_setitem() - self.func_test_setitem() - - def func_test_bump_inplace_version(self): + def test_bump_inplace_version(self): paddle.disable_static() var = paddle.ones(shape=[4, 2, 3], dtype="float32") self.assertEqual(var.inplace_version, 0) @@ -1625,14 +1439,9 @@ class TestVarBaseInplaceVersion(unittest.TestCase): var._bump_inplace_version() self.assertEqual(var.inplace_version, 2) - def test_bump_inplace_version(self): - with _test_eager_guard(): - self.func_test_bump_inplace_version() - self.func_test_bump_inplace_version() - class TestVarBaseSlice(unittest.TestCase): - def func_test_slice(self): + def test_slice(self): paddle.disable_static() np_x = np.random.random((3, 8, 8)) x = paddle.to_tensor(np_x, dtype="float64") @@ -1640,28 +1449,18 @@ class TestVarBaseSlice(unittest.TestCase): actual_x = paddle.to_tensor(actual_x) self.assertEqual(actual_x.numpy().all(), np_x[0:1].all()) - def test_slice(self): - with _test_eager_guard(): - self.func_test_slice() - self.func_test_slice() - class TestVarBaseClear(unittest.TestCase): - def func_test_clear(self): + def test_clear(self): paddle.disable_static() np_x = np.random.random((3, 8, 8)) x = paddle.to_tensor(np_x, dtype="float64") x._clear() self.assertEqual(str(x), "Tensor(Not initialized)") - def test_clear(self): - with _test_eager_guard(): - self.func_test_clear() - self.func_test_clear() - class TestVarBaseOffset(unittest.TestCase): - def func_offset(self): + def test_offset(self): paddle.disable_static() np_x = np.random.random((3, 8, 8)) x = paddle.to_tensor(np_x, dtype="float64") @@ -1670,14 +1469,9 @@ class TestVarBaseOffset(unittest.TestCase): actual_x = paddle.to_tensor(actual_x) self.assertEqual(actual_x._offset(), expected_offset) - def test_offset(self): - with _test_eager_guard(): - self.func_offset() - self.func_offset() - class TestVarBaseShareBufferTo(unittest.TestCase): - def func_test_share_buffer_To(self): + def test_share_buffer_To(self): paddle.disable_static() np_src = np.random.random((3, 8, 8)) src = paddle.to_tensor(np_src, dtype="float64") @@ -1689,11 +1483,6 @@ class TestVarBaseShareBufferTo(unittest.TestCase): src._share_buffer_to(dst) self.assertEqual(src._is_shared_buffer_with(dst), True) - def test_share_buffer_To(self): - with _test_eager_guard(): - self.func_test_share_buffer_To() - self.func_test_share_buffer_To() - class TestVarBaseTo(unittest.TestCase): def func_setUp(self): @@ -1751,15 +1540,12 @@ class TestVarBaseTo(unittest.TestCase): self.assertRaises(AssertionError, self.x._to, blocking=1) def test_to_api(self): - with _test_eager_guard(): - self.func_setUp() - self.func_test_to_api() self.func_setUp() self.func_test_to_api() class TestVarBaseInitVarBaseFromTensorWithDevice(unittest.TestCase): - def func_test_varbase_init(self): + def test_varbase_init(self): paddle.disable_static() t = fluid.Tensor() np_x = np.random.random((3, 8, 8)) @@ -1781,14 +1567,9 @@ class TestVarBaseInitVarBaseFromTensorWithDevice(unittest.TestCase): tmp = fluid.core.eager.Tensor(t, device) self.assertEqual(tmp.numpy().all(), np_x.all()) - def test_varbase_init(self): - with _test_eager_guard(): - self.func_test_varbase_init() - self.func_test_varbase_init() - class TestVarBaseNumel(unittest.TestCase): - def func_test_numel_normal(self): + def test_numel_normal(self): paddle.disable_static() np_x = np.random.random((3, 8, 8)) x = paddle.to_tensor(np_x, dtype="float64") @@ -1796,12 +1577,7 @@ class TestVarBaseNumel(unittest.TestCase): x_expected_numel = np.product((3, 8, 8)) self.assertEqual(x_actual_numel, x_expected_numel) - def test_numel_normal(self): - with _test_eager_guard(): - self.func_test_numel_normal() - self.func_test_numel_normal() - - def func_test_numel_without_holder(self): + def test_numel_without_holder(self): paddle.disable_static() if _in_legacy_dygraph(): x_without_holder = core.VarBase() @@ -1810,14 +1586,9 @@ class TestVarBaseNumel(unittest.TestCase): x_actual_numel = x_without_holder._numel() self.assertEqual(x_actual_numel, 0) - def ttest_numel_without_holder(self): - with _test_eager_guard(): - self.func_test_numel_without_holder() - self.func_test_numel_without_holder() - class TestVarBaseCopyGradientFrom(unittest.TestCase): - def func_test_copy_gradient_from(self): + def test_copy_gradient_from(self): paddle.disable_static() np_x = np.random.random((2, 2)) np_y = np.random.random((2, 2)) @@ -1828,23 +1599,17 @@ class TestVarBaseCopyGradientFrom(unittest.TestCase): x._copy_gradient_from(y) self.assertEqual(x.grad.numpy().all(), np_y.all()) - def test_copy_gradient_from(self): - with _test_eager_guard(): - self.func_test_copy_gradient_from() - self.func_test_copy_gradient_from() - class TestEagerTensorGradNameValue(unittest.TestCase): def test_eager_tensor_grad_name_value(self): - with _test_eager_guard(): - a_np = np.array([2, 3]).astype('float32') - a = paddle.to_tensor(a_np) - a.stop_gradient = False - b = a**2 - self.assertIsNone(a._grad_value()) - b.backward() - # Note, for new dygraph, there are no generated grad name, so we skip the name check. - self.assertIsNotNone(a._grad_value()) + a_np = np.array([2, 3]).astype('float32') + a = paddle.to_tensor(a_np) + a.stop_gradient = False + b = a**2 + self.assertIsNone(a._grad_value()) + b.backward() + # Note, for new dygraph, there are no generated grad name, so we skip the name check. + self.assertIsNotNone(a._grad_value()) if __name__ == '__main__': diff --git a/python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py b/python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py index ab2c091cc0cb4227880cf4dc4b9e2ae3f1484ade..c3e5020507de750958dd0893e5e62322fd4d2fee 100644 --- a/python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py +++ b/python/paddle/fluid/tests/unittests/test_view_op_reuse_allocation.py @@ -17,7 +17,6 @@ import unittest import numpy as np import paddle -from paddle.fluid.framework import _test_eager_guard # NOTE(pangyoki): Tensor View Strategy. @@ -36,7 +35,7 @@ class TestDygraphViewReuseAllocation(unittest.TestCase): def view_api_processing(self, var): return paddle.squeeze(var) - def func_test_view_api(self): + def test_view_api(self): var = paddle.rand(self.input_shape) view_var = self.view_api_processing(var) view_var[0] = 2.0 @@ -47,12 +46,7 @@ class TestDygraphViewReuseAllocation(unittest.TestCase): view_var_numpy = view_var.numpy() np.testing.assert_array_equal(var_numpy, view_var_numpy) - def test_view_api(self): - with _test_eager_guard(): - self.func_test_view_api() - self.func_test_view_api() - - def func_test_forward_version(self): + def test_forward_version(self): var = paddle.rand(self.input_shape) self.assertEqual(var.inplace_version, 0) view_var = self.view_api_processing(var) @@ -69,12 +63,7 @@ class TestDygraphViewReuseAllocation(unittest.TestCase): self.assertEqual(view_var.inplace_version, 2) self.assertEqual(view_var_2.inplace_version, 2) - def test_forward_version(self): - with _test_eager_guard(): - self.func_test_forward_version() - self.func_test_forward_version() - - def func_test_backward_error(self): + def test_backward_error(self): # It raises an error because the inplace operator will result # in incorrect gradient computation. with paddle.fluid.dygraph.guard(): @@ -97,11 +86,6 @@ class TestDygraphViewReuseAllocation(unittest.TestCase): ): loss.backward() - def test_backward_error(self): - with _test_eager_guard(): - self.func_test_backward_error() - self.func_test_backward_error() - class TestUnsqueezeDygraphViewReuseAllocation(TestDygraphViewReuseAllocation): def init_shape(self): diff --git a/python/paddle/fluid/tests/unittests/test_warpctc_op.py b/python/paddle/fluid/tests/unittests/test_warpctc_op.py index c2a272117c8c4457e8ac6a7af11545ca05fa67c3..b8000f3f549bf2465f05fa03212fae3673e63ba7 100644 --- a/python/paddle/fluid/tests/unittests/test_warpctc_op.py +++ b/python/paddle/fluid/tests/unittests/test_warpctc_op.py @@ -23,7 +23,6 @@ import paddle import paddle.fluid.core as core import paddle.nn.functional as F from paddle.fluid import Program, program_guard -from paddle.fluid.framework import _test_eager_guard paddle.enable_static() @@ -718,8 +717,6 @@ class TestCTCLossAPICase(unittest.TestCase): loss_pd_sum, loss_np_sum, rtol=1e-05, atol=1 ) - with _test_eager_guard(): - test_functinal_api() test_functinal_api()