未验证 提交 f55f601e 编写于 作者: C co63oc 提交者: GitHub

Fix typos, test=document_fix (#56005)

上级 bbafed4c
......@@ -397,7 +397,7 @@ class AmpTestBase(unittest.TestCase):
self.assertEqual(
actual_value,
expected_value,
f"[{debug_info}] The number of bf16 calls of operator < {op_type} > is expected to be {expected_value}, but recieved {actual_value}.",
f"[{debug_info}] The number of bf16 calls of operator < {op_type} > is expected to be {expected_value}, but received {actual_value}.",
)
for op_type, expected_value in expected_fp16_calls.items():
# print(f"[FP16] op_type={op_type}, value={value}")
......@@ -408,7 +408,7 @@ class AmpTestBase(unittest.TestCase):
self.assertEqual(
actual_value,
expected_value,
f"[debug_info] The number of fp16 calls of operator < {op_type} > is expected to be {expected_value}, but recieved {actual_value}.",
f"[debug_info] The number of fp16 calls of operator < {op_type} > is expected to be {expected_value}, but received {actual_value}.",
)
def run_program(
......
......@@ -46,7 +46,7 @@ class TestStaticMasterGradProgramFP16(AmpTestBase):
self.assertEqual(
actual_num_mp,
expected_num_mp,
f"The number of optimizers with multi_precison = True is expected to be {expected_num_mp}, but recieved {actual_num_mp}.",
f"The number of optimizers with multi_precison = True is expected to be {expected_num_mp}, but received {actual_num_mp}.",
)
def amp_fp16_o2(self, use_master_grad):
......@@ -189,13 +189,13 @@ class TestMasterGradAccuracy(AmpTestBase):
self.assertNotEqual(
losses_o1,
losses_o2_no_master_grad,
f"dtype: {dtype}, loss of o1 and o2-wo-master_grad should not be equal, but recieved loss o1: {losses_o1}, loss o2: {losses_o2_no_master_grad}",
f"dtype: {dtype}, loss of o1 and o2-wo-master_grad should not be equal, but received loss o1: {losses_o1}, loss o2: {losses_o2_no_master_grad}",
)
self.assertEqual(
losses_o1,
losses_o2_master_grad,
f"dtype: {dtype}, loss of o1 and o2-w-master_grad should be equal, but recieved loss o1: {losses_o1}, loss o2: {losses_o2_master_grad}",
f"dtype: {dtype}, loss of o1 and o2-w-master_grad should be equal, but received loss o1: {losses_o1}, loss o2: {losses_o2_master_grad}",
)
......
......@@ -212,7 +212,7 @@ class TestProgramBF16(AmpTestBase):
self.assertEqual(
actual_num_mp,
expected_num_mp,
f"The number of optimizers with multi_precison = True is expected to be {expected_num_mp}, but recieved {actual_num_mp}.",
f"The number of optimizers with multi_precison = True is expected to be {expected_num_mp}, but received {actual_num_mp}.",
)
def test_amp_bf16_o1(self):
......@@ -302,7 +302,7 @@ class TestStaticBF16(AmpTestBase):
self.assertEqual(
losses_o1,
losses_o2,
f"loss of o1 and o2 should be equal, but recieved loss o1: {losses_o1}, loss o2: {losses_o2}",
f"loss of o1 and o2 should be equal, but received loss o1: {losses_o1}, loss o2: {losses_o2}",
)
......
......@@ -96,7 +96,7 @@ class TestTensorChecker(unittest.TestCase):
self.assertEqual(
1,
num_nan,
f"Expected num_nan to be 1, but recieved {num_nan}, place={place}.",
f"Expected num_nan to be 1, but received {num_nan}, place={place}.",
)
else:
self.assertEqual(
......@@ -111,7 +111,7 @@ class TestTensorChecker(unittest.TestCase):
self.assertEqual(
0,
num_nan,
f"Expected num_nan to be 1, but recieved {num_nan}, place={place}.",
f"Expected num_nan to be 1, but received {num_nan}, place={place}.",
)
paddle.amp.debugging.disable_tensor_checker()
......
......@@ -83,7 +83,7 @@ class TestNanInf(TestNanInfBase):
self.assertEqual(
actual_value,
expected_value,
f"The number of operator < {op_type} > is expected to be {expected_value}, but recieved {actual_value}.",
f"The number of operator < {op_type} > is expected to be {expected_value}, but received {actual_value}.",
)
print("")
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册