From ad01eccdecba99793ec9d24a18b43f5d5ee0a673 Mon Sep 17 00:00:00 2001 From: wuyefeilin <30919197+wuyefeilin@users.noreply.github.com> Date: Wed, 29 Mar 2023 16:37:30 +0800 Subject: [PATCH] [AMP OP&Test] Add fp16/bf16 to clip op (#52158) * add fp16/bf16 to clip op * fix as reviewed * update test_clip_op.py * update test_clip_op.py --- .../fluid/tests/unittests/test_clip_op.py | 131 +++++++++++++++++- python/paddle/tensor/math.py | 9 +- 2 files changed, 134 insertions(+), 6 deletions(-) diff --git a/python/paddle/fluid/tests/unittests/test_clip_op.py b/python/paddle/fluid/tests/unittests/test_clip_op.py index 979ce6e7cc9..eca97092a64 100644 --- a/python/paddle/fluid/tests/unittests/test_clip_op.py +++ b/python/paddle/fluid/tests/unittests/test_clip_op.py @@ -15,11 +15,11 @@ import unittest import numpy as np -from eager_op_test import OpTest +from eager_op_test import OpTest, convert_float_to_uint16 import paddle from paddle import fluid -from paddle.fluid import Program, program_guard +from paddle.fluid import Program, core, program_guard class TestClipOp(OpTest): @@ -111,7 +111,31 @@ class TestCase5(TestClipOp): self.min = 0.5 -class TestCase6(TestClipOp): +class TestFP16Case1(TestClipOp): + def initTestCase(self): + self.dtype = np.float16 + self.shape = (8, 16, 8) + self.max = 0.7 + self.min = 0.0 + + +class TestFP16Case2(TestClipOp): + def initTestCase(self): + self.dtype = np.float16 + self.shape = (8, 16) + self.max = 1.0 + self.min = 0.0 + + +class TestFP16Case3(TestClipOp): + def initTestCase(self): + self.dtype = np.float16 + self.shape = (4, 8, 16) + self.max = 0.7 + self.min = 0.2 + + +class TestFP16Case4(TestClipOp): def initTestCase(self): self.dtype = np.float16 self.shape = (4, 8, 8) @@ -121,6 +145,107 @@ class TestCase6(TestClipOp): self.inputs['Min'] = np.array([0.3]).astype(self.dtype) +class TestFP16Case5(TestClipOp): + def initTestCase(self): + self.dtype = np.float16 + self.shape = (4, 8, 16) + self.max = 0.5 + self.min = 0.5 + + +@unittest.skipIf( + not core.is_compiled_with_cuda() + or not core.is_bfloat16_supported(core.CUDAPlace(0)), + "core is not compiled with CUDA and not support the bfloat16", +) +class TestClipBF16Op(OpTest): + def setUp(self): + self.max_relative_error = 0.006 + self.python_api = paddle.clip + + self.inputs = {} + self.initTestCase() + + self.op_type = "clip" + self.attrs = {} + self.attrs['min'] = self.min + self.attrs['max'] = self.max + if 'Min' in self.inputs: + min_v = self.inputs['Min'] + else: + min_v = self.attrs['min'] + + if 'Max' in self.inputs: + max_v = self.inputs['Max'] + else: + max_v = self.attrs['max'] + + input = np.random.random(self.shape).astype(np.float32) + input[np.abs(input - min_v) < self.max_relative_error] = 0.5 + input[np.abs(input - max_v) < self.max_relative_error] = 0.5 + self.inputs['X'] = convert_float_to_uint16(input) + out = np.clip(input, min_v, max_v) + self.outputs = {'Out': convert_float_to_uint16(out)} + + def test_check_output(self): + if paddle.is_compiled_with_cuda(): + place = paddle.CUDAPlace(0) + paddle.enable_static() + self.check_output_with_place(place) + paddle.disable_static() + + def test_check_grad_normal(self): + if paddle.is_compiled_with_cuda(): + place = paddle.CUDAPlace(0) + paddle.enable_static() + self.check_grad_with_place(place, ['X'], 'Out') + paddle.disable_static() + + def initTestCase(self): + self.shape = (4, 10, 10) + self.max = 0.8 + self.min = 0.3 + self.inputs['Max'] = np.array([0.8]).astype(np.float32) + self.inputs['Min'] = np.array([0.1]).astype(np.float32) + + +class TestBF16Case1(TestClipBF16Op): + def initTestCase(self): + self.shape = (8, 16, 8) + self.max = 0.7 + self.min = 0.0 + + +class TestBF16Case2(TestClipBF16Op): + def initTestCase(self): + self.shape = (8, 16) + self.max = 1.0 + self.min = 0.0 + + +class TestBF16Case3(TestClipBF16Op): + def initTestCase(self): + self.shape = (4, 8, 16) + self.max = 0.7 + self.min = 0.2 + + +class TestBF16Case4(TestClipBF16Op): + def initTestCase(self): + self.shape = (4, 8, 8) + self.max = 0.7 + self.min = 0.2 + self.inputs['Max'] = np.array([0.8]).astype(np.float32) + self.inputs['Min'] = np.array([0.3]).astype(np.float32) + + +class TestBF16Case5(TestClipBF16Op): + def initTestCase(self): + self.shape = (4, 8, 16) + self.max = 0.5 + self.min = 0.5 + + class TestClipOpError(unittest.TestCase): def test_errors(self): paddle.enable_static() diff --git a/python/paddle/tensor/math.py b/python/paddle/tensor/math.py index cd4748c3c0a..8d889c964c4 100644 --- a/python/paddle/tensor/math.py +++ b/python/paddle/tensor/math.py @@ -2885,7 +2885,7 @@ def clip(x, min=None, max=None, name=None): check_dtype( min.dtype, 'min', - ['float16', 'float32', 'float64', 'int32'], + ['float16', 'float32', 'float64', 'int32', 'uint16'], 'clip', '(When the type of min in clip is Variable.)', ) @@ -2895,13 +2895,16 @@ def clip(x, min=None, max=None, name=None): check_dtype( max.dtype, 'max', - ['float16', 'float32', 'float64', 'int32'], + ['float16', 'float32', 'float64', 'int32', 'uint16'], 'clip', '(When the type of max in clip is Variable.)', ) check_variable_and_dtype( - x, 'x', ['float16', 'float32', 'float64', 'int32', 'int64'], 'clip' + x, + 'x', + ['float16', 'float32', 'float64', 'int32', 'int64', 'uint16'], + 'clip', ) inputs = {'X': x} -- GitLab