未验证 提交 cb34ee0f 编写于 作者: 姜永久 提交者: GitHub

rm eager guard test (#49245)

* rm eager guard test

* retain grad for xpu test
上级 ca4155c8
...@@ -38,7 +38,6 @@ from paddle.nn.quant.quant_layers import ( ...@@ -38,7 +38,6 @@ from paddle.nn.quant.quant_layers import (
QuantizedConv2D, QuantizedConv2D,
QuantizedMatmul, QuantizedMatmul,
) )
from paddle.fluid.framework import _test_eager_guard
from imperative_test_utils import fix_model_dict from imperative_test_utils import fix_model_dict
paddle.enable_static() paddle.enable_static()
......
...@@ -21,7 +21,6 @@ import numpy as np ...@@ -21,7 +21,6 @@ import numpy as np
import paddle import paddle
import paddle.static as static import paddle.static as static
from paddle.fluid.framework import _test_eager_guard
from paddle.utils.cpp_extension.extension_utils import run_cmd from paddle.utils.cpp_extension.extension_utils import run_cmd
from paddle.vision.transforms import Compose, Normalize from paddle.vision.transforms import Compose, Normalize
...@@ -143,6 +142,9 @@ def custom_relu_static_inference(func, device, np_data, np_label, path_prefix): ...@@ -143,6 +142,9 @@ def custom_relu_static_inference(func, device, np_data, np_label, path_prefix):
def custom_relu_double_grad_dynamic(func, device, dtype, np_x, use_func=True): def custom_relu_double_grad_dynamic(func, device, dtype, np_x, use_func=True):
import paddle.fluid as fluid
fluid.set_flags({"FLAGS_retain_grad_for_all_tensor": True})
paddle.set_device(device) paddle.set_device(device)
t = paddle.to_tensor(np_x, dtype=dtype, stop_gradient=False) t = paddle.to_tensor(np_x, dtype=dtype, stop_gradient=False)
...@@ -158,6 +160,7 @@ def custom_relu_double_grad_dynamic(func, device, dtype, np_x, use_func=True): ...@@ -158,6 +160,7 @@ def custom_relu_double_grad_dynamic(func, device, dtype, np_x, use_func=True):
assert dx[0].grad is not None assert dx[0].grad is not None
return dx[0].numpy(), dx[0].grad.numpy() return dx[0].numpy(), dx[0].grad.numpy()
fluid.set_flags({"FLAGS_retain_grad_for_all_tensor": False})
class TestNewCustomOpSetUpInstall(unittest.TestCase): class TestNewCustomOpSetUpInstall(unittest.TestCase):
...@@ -228,7 +231,7 @@ class TestNewCustomOpSetUpInstall(unittest.TestCase): ...@@ -228,7 +231,7 @@ class TestNewCustomOpSetUpInstall(unittest.TestCase):
), ),
) )
def func_dynamic(self): def test_dynamic(self):
for device in self.devices: for device in self.devices:
for dtype in self.dtypes: for dtype in self.dtypes:
x = np.random.uniform(-1, 1, [4, 8]).astype(dtype) x = np.random.uniform(-1, 1, [4, 8]).astype(dtype)
...@@ -253,11 +256,6 @@ class TestNewCustomOpSetUpInstall(unittest.TestCase): ...@@ -253,11 +256,6 @@ class TestNewCustomOpSetUpInstall(unittest.TestCase):
), ),
) )
def test_dynamic(self):
with _test_eager_guard():
self.func_dynamic()
self.func_dynamic()
def test_static_save_and_load_inference_model(self): def test_static_save_and_load_inference_model(self):
paddle.enable_static() paddle.enable_static()
np_data = np.random.random((1, 1, 28, 28)).astype("float32") np_data = np.random.random((1, 1, 28, 28)).astype("float32")
......
...@@ -18,7 +18,6 @@ import numpy as np ...@@ -18,7 +18,6 @@ import numpy as np
import paddle import paddle
from paddle import nn from paddle import nn
from paddle.distributed.sharding import group_sharded_parallel from paddle.distributed.sharding import group_sharded_parallel
from paddle.fluid.framework import _test_eager_guard
paddle.seed(2022) paddle.seed(2022)
np.random.seed(2022) np.random.seed(2022)
...@@ -174,5 +173,4 @@ def test_sharding_api(): ...@@ -174,5 +173,4 @@ def test_sharding_api():
if __name__ == '__main__': if __name__ == '__main__':
with _test_eager_guard():
test_sharding_api() test_sharding_api()
...@@ -20,7 +20,6 @@ import numpy as np ...@@ -20,7 +20,6 @@ import numpy as np
from test_dist_base import RUN_STEP from test_dist_base import RUN_STEP
import paddle import paddle
from paddle.fluid.framework import _test_eager_guard
# NOTE: compatible TestParallelDyGraphRunnerBase args # NOTE: compatible TestParallelDyGraphRunnerBase args
...@@ -54,10 +53,6 @@ class TestDistSpawnRunner(unittest.TestCase): ...@@ -54,10 +53,6 @@ class TestDistSpawnRunner(unittest.TestCase):
return result_list return result_list
def check_dist_result_with_spawn(self, test_class, delta=1e-3): def check_dist_result_with_spawn(self, test_class, delta=1e-3):
with _test_eager_guard():
self.check_dist_result_with_spawn_func(
test_class=test_class, delta=delta
)
self.check_dist_result_with_spawn_func( self.check_dist_result_with_spawn_func(
test_class=test_class, delta=delta test_class=test_class, delta=delta
) )
......
...@@ -19,7 +19,6 @@ import numpy as np ...@@ -19,7 +19,6 @@ import numpy as np
import paddle import paddle
import paddle.fluid as fluid import paddle.fluid as fluid
from paddle.fluid.framework import _test_eager_guard
class TestMathOpPatchesVarBase(unittest.TestCase): class TestMathOpPatchesVarBase(unittest.TestCase):
...@@ -27,7 +26,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -27,7 +26,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
self.shape = [10, 1024] self.shape = [10, 1024]
self.dtype = np.float32 self.dtype = np.float32
def func_test_add(self): def test_add(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -36,12 +35,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -36,12 +35,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a + b res = a + b
np.testing.assert_array_equal(res.numpy(), a_np + b_np) np.testing.assert_array_equal(res.numpy(), a_np + b_np)
def test_add(self): def test_sub(self):
with _test_eager_guard():
self.func_test_add()
self.func_test_add()
def func_test_sub(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -50,12 +44,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -50,12 +44,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a - b res = a - b
np.testing.assert_array_equal(res.numpy(), a_np - b_np) np.testing.assert_array_equal(res.numpy(), a_np - b_np)
def test_sub(self): def test_mul(self):
with _test_eager_guard():
self.func_test_sub()
self.func_test_sub()
def func_test_mul(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -64,12 +53,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -64,12 +53,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a * b res = a * b
np.testing.assert_array_equal(res.numpy(), a_np * b_np) np.testing.assert_array_equal(res.numpy(), a_np * b_np)
def test_mul(self): def test_div(self):
with _test_eager_guard():
self.func_test_mul()
self.func_test_mul()
def func_test_div(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -79,12 +63,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -79,12 +63,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
# NOTE: Not sure why array_equal fails on windows, allclose is acceptable # NOTE: Not sure why array_equal fails on windows, allclose is acceptable
np.testing.assert_allclose(res.numpy(), a_np / b_np, rtol=1e-05) np.testing.assert_allclose(res.numpy(), a_np / b_np, rtol=1e-05)
def test_div(self): def test_add_scalar(self):
with _test_eager_guard():
self.func_test_div()
self.func_test_div()
def func_test_add_scalar(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -92,12 +71,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -92,12 +71,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a + b res = a + b
np.testing.assert_array_equal(res.numpy(), a_np + b) np.testing.assert_array_equal(res.numpy(), a_np + b)
def test_add_scalar(self): def test_add_scalar_reverse(self):
with _test_eager_guard():
self.func_test_add_scalar()
self.func_test_add_scalar()
def func_test_add_scalar_reverse(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -105,12 +79,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -105,12 +79,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = b + a res = b + a
np.testing.assert_array_equal(res.numpy(), b + a_np) np.testing.assert_array_equal(res.numpy(), b + a_np)
def test_add_scalar_reverse(self): def test_sub_scalar(self):
with _test_eager_guard():
self.func_test_add_scalar_reverse()
self.func_test_add_scalar_reverse()
def func_test_sub_scalar(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -118,12 +87,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -118,12 +87,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a - b res = a - b
np.testing.assert_array_equal(res.numpy(), a_np - b) np.testing.assert_array_equal(res.numpy(), a_np - b)
def test_sub_scalar(self): def test_sub_scalar_reverse(self):
with _test_eager_guard():
self.func_test_sub_scalar()
self.func_test_sub_scalar()
def func_test_sub_scalar_reverse(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -131,12 +95,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -131,12 +95,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = b - a res = b - a
np.testing.assert_array_equal(res.numpy(), b - a_np) np.testing.assert_array_equal(res.numpy(), b - a_np)
def test_sub_scalar_reverse(self): def test_mul_scalar(self):
with _test_eager_guard():
self.func_test_sub_scalar_reverse()
self.func_test_sub_scalar_reverse()
def func_test_mul_scalar(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -144,13 +103,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -144,13 +103,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a * b res = a * b
np.testing.assert_array_equal(res.numpy(), a_np * b) np.testing.assert_array_equal(res.numpy(), a_np * b)
def test_mul_scalar(self):
with _test_eager_guard():
self.func_test_mul_scalar()
self.func_test_mul_scalar()
# div_scalar, not equal # div_scalar, not equal
def func_test_div_scalar(self): def test_div_scalar(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -158,13 +112,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -158,13 +112,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a / b res = a / b
np.testing.assert_allclose(res.numpy(), a_np / b, rtol=1e-05) np.testing.assert_allclose(res.numpy(), a_np / b, rtol=1e-05)
def test_div_scalar(self):
with _test_eager_guard():
self.func_test_div_scalar()
self.func_test_div_scalar()
# pow of float type, not equal # pow of float type, not equal
def func_test_pow(self): def test_pow(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -173,12 +122,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -173,12 +122,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a**b res = a**b
np.testing.assert_allclose(res.numpy(), a_np**b_np, rtol=1e-05) np.testing.assert_allclose(res.numpy(), a_np**b_np, rtol=1e-05)
def test_pow(self): def test_floor_div(self):
with _test_eager_guard():
self.func_test_pow()
self.func_test_pow()
def func_test_floor_div(self):
a_np = np.random.randint(1, 100, size=self.shape) a_np = np.random.randint(1, 100, size=self.shape)
b_np = np.random.randint(1, 100, size=self.shape) b_np = np.random.randint(1, 100, size=self.shape)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -187,12 +131,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -187,12 +131,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a // b res = a // b
np.testing.assert_array_equal(res.numpy(), a_np // b_np) np.testing.assert_array_equal(res.numpy(), a_np // b_np)
def test_floor_div(self): def test_mod(self):
with _test_eager_guard():
self.func_test_floor_div()
self.func_test_floor_div()
def func_test_mod(self):
a_np = np.random.randint(1, 100, size=self.shape) a_np = np.random.randint(1, 100, size=self.shape)
b_np = np.random.randint(1, 100, size=self.shape) b_np = np.random.randint(1, 100, size=self.shape)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -201,13 +140,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -201,13 +140,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a % b res = a % b
np.testing.assert_array_equal(res.numpy(), a_np % b_np) np.testing.assert_array_equal(res.numpy(), a_np % b_np)
def test_mod(self):
with _test_eager_guard():
self.func_test_mod()
self.func_test_mod()
# for bitwise and/or/xor/not # for bitwise and/or/xor/not
def func_test_bitwise(self): def test_bitwise(self):
paddle.disable_static() paddle.disable_static()
x_np = np.random.randint(-100, 100, [2, 3, 5]) x_np = np.random.randint(-100, 100, [2, 3, 5])
...@@ -231,13 +165,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -231,13 +165,8 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
out = ~x out = ~x
np.testing.assert_array_equal(out.numpy(), out_np) np.testing.assert_array_equal(out.numpy(), out_np)
def test_bitwise(self):
with _test_eager_guard():
self.func_test_bitwise()
self.func_test_bitwise()
# for logical compare # for logical compare
def func_test_equal(self): def test_equal(self):
a_np = np.asarray([1, 2, 3, 4, 5]) a_np = np.asarray([1, 2, 3, 4, 5])
b_np = np.asarray([1, 2, 3, 4, 5]) b_np = np.asarray([1, 2, 3, 4, 5])
c_np = np.asarray([1, 2, 2, 4, 5]) c_np = np.asarray([1, 2, 2, 4, 5])
...@@ -250,12 +179,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -250,12 +179,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
np.testing.assert_array_equal(res1.numpy(), a_np == b_np) np.testing.assert_array_equal(res1.numpy(), a_np == b_np)
np.testing.assert_array_equal(res2.numpy(), a_np == c_np) np.testing.assert_array_equal(res2.numpy(), a_np == c_np)
def test_equal(self): def test_not_equal(self):
with _test_eager_guard():
self.func_test_equal()
self.func_test_equal()
def func_test_not_equal(self):
a_np = np.asarray([1, 2, 3, 4, 5]) a_np = np.asarray([1, 2, 3, 4, 5])
b_np = np.asarray([1, 2, 3, 4, 5]) b_np = np.asarray([1, 2, 3, 4, 5])
c_np = np.asarray([1, 2, 2, 4, 5]) c_np = np.asarray([1, 2, 2, 4, 5])
...@@ -268,12 +192,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -268,12 +192,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
np.testing.assert_array_equal(res1.numpy(), a_np != b_np) np.testing.assert_array_equal(res1.numpy(), a_np != b_np)
np.testing.assert_array_equal(res2.numpy(), a_np != c_np) np.testing.assert_array_equal(res2.numpy(), a_np != c_np)
def test_not_equal(self): def test_less_than(self):
with _test_eager_guard():
self.func_test_not_equal()
self.func_test_not_equal()
def func_test_less_than(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -282,12 +201,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -282,12 +201,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a < b res = a < b
np.testing.assert_array_equal(res.numpy(), a_np < b_np) np.testing.assert_array_equal(res.numpy(), a_np < b_np)
def test_less_than(self): def test_less_equal(self):
with _test_eager_guard():
self.func_test_less_than()
self.func_test_less_than()
def func_test_less_equal(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -296,12 +210,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -296,12 +210,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a <= b res = a <= b
np.testing.assert_array_equal(res.numpy(), a_np <= b_np) np.testing.assert_array_equal(res.numpy(), a_np <= b_np)
def test_less_equal(self): def test_greater_than(self):
with _test_eager_guard():
self.func_test_less_equal()
self.func_test_less_equal()
def func_test_greater_than(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -310,12 +219,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -310,12 +219,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a > b res = a > b
np.testing.assert_array_equal(res.numpy(), a_np > b_np) np.testing.assert_array_equal(res.numpy(), a_np > b_np)
def test_greater_than(self): def test_greater_equal(self):
with _test_eager_guard():
self.func_test_greater_than()
self.func_test_greater_than()
def func_test_greater_equal(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
b_np = np.random.random(self.shape).astype(self.dtype) b_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -324,47 +228,27 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -324,47 +228,27 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a >= b res = a >= b
np.testing.assert_array_equal(res.numpy(), a_np >= b_np) np.testing.assert_array_equal(res.numpy(), a_np >= b_np)
def test_greater_equal(self): def test_neg(self):
with _test_eager_guard():
self.func_test_greater_equal()
self.func_test_greater_equal()
def func_test_neg(self):
a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype) a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
res = -a res = -a
np.testing.assert_array_equal(res.numpy(), -a_np) np.testing.assert_array_equal(res.numpy(), -a_np)
def test_neg(self): def test_float_int_long(self):
with _test_eager_guard():
self.func_test_neg()
self.func_test_neg()
def func_test_float_int_long(self):
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(np.array([100.1])) a = fluid.dygraph.to_variable(np.array([100.1]))
self.assertTrue(float(a) == 100.1) self.assertTrue(float(a) == 100.1)
self.assertTrue(int(a) == 100) self.assertTrue(int(a) == 100)
self.assertTrue(int(a) == 100) self.assertTrue(int(a) == 100)
def test_float_int_long(self): def test_len(self):
with _test_eager_guard():
self.func_test_float_int_long()
self.func_test_float_int_long()
def func_test_len(self):
a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype) a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
self.assertTrue(len(a) == 10) self.assertTrue(len(a) == 10)
def test_len(self): def test_index(self):
with _test_eager_guard():
self.func_test_len()
self.func_test_len()
def func_test_index(self):
with fluid.dygraph.guard(): with fluid.dygraph.guard():
var1 = fluid.dygraph.to_variable(np.array([2])) var1 = fluid.dygraph.to_variable(np.array([2]))
i_tmp = 0 i_tmp = 0
...@@ -376,12 +260,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -376,12 +260,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
str1 = "just test" str1 = "just test"
self.assertTrue(str1[var1] == 's') self.assertTrue(str1[var1] == 's')
def test_index(self): def test_np_left_mul(self):
with _test_eager_guard():
self.func_test_index()
self.func_test_index()
def func_test_np_left_mul(self):
with fluid.dygraph.guard(): with fluid.dygraph.guard():
t = np.sqrt(2.0 * np.pi) t = np.sqrt(2.0 * np.pi)
x = paddle.ones((2, 2), dtype="float32") x = paddle.ones((2, 2), dtype="float32")
...@@ -394,12 +273,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -394,12 +273,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
atol=0.0, atol=0.0,
) )
def test_np_left_mul(self): def test_add_different_dtype(self):
with _test_eager_guard():
self.func_test_np_left_mul()
self.func_test_np_left_mul()
def func_test_add_different_dtype(self):
a_np = np.random.random(self.shape).astype(np.float32) a_np = np.random.random(self.shape).astype(np.float32)
b_np = np.random.random(self.shape).astype(np.float16) b_np = np.random.random(self.shape).astype(np.float16)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -408,12 +282,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -408,12 +282,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a + b res = a + b
np.testing.assert_array_equal(res.numpy(), a_np + b_np) np.testing.assert_array_equal(res.numpy(), a_np + b_np)
def test_add_different_dtype(self): def test_floordiv_different_dtype(self):
with _test_eager_guard():
self.func_test_add_different_dtype()
self.func_test_add_different_dtype()
def func_test_floordiv_different_dtype(self):
a_np = np.full(self.shape, 10, np.int64) a_np = np.full(self.shape, 10, np.int64)
b_np = np.full(self.shape, 2, np.int32) b_np = np.full(self.shape, 2, np.int32)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -422,12 +291,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -422,12 +291,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
res = a // b res = a // b
np.testing.assert_array_equal(res.numpy(), a_np // b_np) np.testing.assert_array_equal(res.numpy(), a_np // b_np)
def test_floordiv_different_dtype(self): def test_astype(self):
with _test_eager_guard():
self.func_test_floordiv_different_dtype()
self.func_test_floordiv_different_dtype()
def func_test_astype(self):
a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype) a_np = np.random.uniform(-1, 1, self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
...@@ -441,12 +305,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -441,12 +305,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
np.testing.assert_array_equal(res1.numpy(), res2.numpy()) np.testing.assert_array_equal(res1.numpy(), res2.numpy())
np.testing.assert_array_equal(res1.numpy(), res3.numpy()) np.testing.assert_array_equal(res1.numpy(), res3.numpy())
def test_astype(self): def test_conpare_op_broadcast(self):
with _test_eager_guard():
self.func_test_astype()
self.func_test_astype()
def func_test_conpare_op_broadcast(self):
a_np = np.random.uniform(-1, 1, [10, 1, 10]).astype(self.dtype) a_np = np.random.uniform(-1, 1, [10, 1, 10]).astype(self.dtype)
b_np = np.random.uniform(-1, 1, [1, 1, 10]).astype(self.dtype) b_np = np.random.uniform(-1, 1, [1, 1, 10]).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
...@@ -456,12 +315,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -456,12 +315,7 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
self.assertEqual((a != b).dtype, fluid.core.VarDesc.VarType.BOOL) self.assertEqual((a != b).dtype, fluid.core.VarDesc.VarType.BOOL)
np.testing.assert_array_equal((a != b).numpy(), a_np != b_np) np.testing.assert_array_equal((a != b).numpy(), a_np != b_np)
def test_conpare_op_broadcast(self): def test_tensor_patch_method(self):
with _test_eager_guard():
self.func_test_conpare_op_broadcast()
self.func_test_conpare_op_broadcast()
def func_test_tensor_patch_method(self):
paddle.disable_static() paddle.disable_static()
x_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype) x_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype)
y_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype) y_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype)
...@@ -740,24 +594,14 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -740,24 +594,14 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
self.assertTrue(inspect.ismethod(a.std)) self.assertTrue(inspect.ismethod(a.std))
self.assertTrue(inspect.ismethod(a.numel)) self.assertTrue(inspect.ismethod(a.numel))
def test_tensor_patch_method(self): def test_complex_scalar(self):
with _test_eager_guard():
self.func_test_tensor_patch_method()
self.func_test_tensor_patch_method()
def func_test_complex_scalar(self):
a_np = np.random.random(self.shape).astype(self.dtype) a_np = np.random.random(self.shape).astype(self.dtype)
with fluid.dygraph.guard(): with fluid.dygraph.guard():
a = fluid.dygraph.to_variable(a_np) a = fluid.dygraph.to_variable(a_np)
res = 1j * a res = 1j * a
np.testing.assert_array_equal(res.numpy(), 1j * a_np) np.testing.assert_array_equal(res.numpy(), 1j * a_np)
def test_complex_scalar(self): def test_matmul(self):
with _test_eager_guard():
self.func_test_complex_scalar()
self.func_test_complex_scalar()
def func_test_matmul(self):
x_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype) x_np = np.random.uniform(-1, 1, [2, 3]).astype(self.dtype)
y_np = np.random.uniform(-1, 1, [3, 2]).astype(self.dtype) y_np = np.random.uniform(-1, 1, [3, 2]).astype(self.dtype)
except_out = x_np @ y_np except_out = x_np @ y_np
...@@ -768,11 +612,6 @@ class TestMathOpPatchesVarBase(unittest.TestCase): ...@@ -768,11 +612,6 @@ class TestMathOpPatchesVarBase(unittest.TestCase):
out = x @ y out = x @ y
np.testing.assert_allclose(out.numpy(), except_out, atol=1e-03) np.testing.assert_allclose(out.numpy(), except_out, atol=1e-03)
def test_matmul(self):
with _test_eager_guard():
self.func_test_matmul()
self.func_test_matmul()
if __name__ == '__main__': if __name__ == '__main__':
unittest.main() unittest.main()
...@@ -28,7 +28,6 @@ from xpu.get_test_cover_info import ( ...@@ -28,7 +28,6 @@ from xpu.get_test_cover_info import (
import paddle import paddle
import paddle.fluid as fluid import paddle.fluid as fluid
from paddle.fluid import Program, program_guard from paddle.fluid import Program, program_guard
from paddle.fluid.framework import _test_eager_guard
paddle.enable_static() paddle.enable_static()
...@@ -305,8 +304,6 @@ class XPUTestDiagV2Op(XPUOpTestWrapper): ...@@ -305,8 +304,6 @@ class XPUTestDiagV2Op(XPUOpTestWrapper):
def test_xpu(self): def test_xpu(self):
paddle.disable_static(place=paddle.fluid.XPUPlace(0)) paddle.disable_static(place=paddle.fluid.XPUPlace(0))
self.run_imperative() self.run_imperative()
with _test_eager_guard():
self.run_imperative()
paddle.enable_static() paddle.enable_static()
......
...@@ -28,7 +28,6 @@ from xpu.get_test_cover_info import ( ...@@ -28,7 +28,6 @@ from xpu.get_test_cover_info import (
) )
import paddle import paddle
from paddle.fluid.framework import _test_eager_guard
class XPUTestSetValueOp(XPUOpTestWrapper): class XPUTestSetValueOp(XPUOpTestWrapper):
...@@ -66,7 +65,7 @@ class XPUTestSetValueOp(XPUOpTestWrapper): ...@@ -66,7 +65,7 @@ class XPUTestSetValueOp(XPUOpTestWrapper):
paddle.disable_static() paddle.disable_static()
return out return out
def func_test_api(self): def test_api(self):
static_out = self._run_static() static_out = self._run_static()
self._get_answer() self._get_answer()
...@@ -78,11 +77,6 @@ class XPUTestSetValueOp(XPUOpTestWrapper): ...@@ -78,11 +77,6 @@ class XPUTestSetValueOp(XPUOpTestWrapper):
msg=error_msg.format("static", self.data, static_out), msg=error_msg.format("static", self.data, static_out),
) )
def test_api(self):
with _test_eager_guard():
self.func_test_api()
self.func_test_api()
support_types = get_xpu_op_support_types('set_value') support_types = get_xpu_op_support_types('set_value')
for stype in support_types: for stype in support_types:
......
...@@ -17,7 +17,6 @@ import unittest ...@@ -17,7 +17,6 @@ import unittest
import paddle import paddle
import paddle.vision.transforms as T import paddle.vision.transforms as T
from paddle.fluid.framework import _test_eager_guard
from paddle.static import InputSpec from paddle.static import InputSpec
from paddle.vision.datasets import MNIST from paddle.vision.datasets import MNIST
...@@ -31,7 +30,7 @@ class TestWandbCallbacks(unittest.TestCase): ...@@ -31,7 +30,7 @@ class TestWandbCallbacks(unittest.TestCase):
def setUp(self): def setUp(self):
self.save_dir = tempfile.mkdtemp() self.save_dir = tempfile.mkdtemp()
def func_wandb_callback(self): def test_wandb_callback(self):
inputs = [InputSpec([-1, 1, 28, 28], 'float32', 'image')] inputs = [InputSpec([-1, 1, 28, 28], 'float32', 'image')]
labels = [InputSpec([None, 1], 'int64', 'label')] labels = [InputSpec([None, 1], 'int64', 'label')]
...@@ -59,11 +58,6 @@ class TestWandbCallbacks(unittest.TestCase): ...@@ -59,11 +58,6 @@ class TestWandbCallbacks(unittest.TestCase):
train_dataset, eval_dataset, batch_size=64, callbacks=callback train_dataset, eval_dataset, batch_size=64, callbacks=callback
) )
def test_wandb_callback(self):
with _test_eager_guard():
self.func_wandb_callback()
self.func_wandb_callback()
if __name__ == '__main__': if __name__ == '__main__':
unittest.main() unittest.main()
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册