test_matmul_op_xpu.py 12.8 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15
#   Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import unittest
16

17
import numpy as np
R
RedContritio 已提交
18
from get_test_cover_info import (
19
    XPUOpTestWrapper,
20 21 22
    create_test_class,
    get_xpu_op_support_types,
)
R
RedContritio 已提交
23
from op_test_xpu import XPUOpTest
24

25
import paddle
26
from paddle import fluid
27

28 29 30 31 32 33 34 35 36 37 38

def reference_matmul(X, Y, transpose_X=False, transpose_Y=False):
    """Reference forward implementation using np.matmul."""
    # np.matmul does not support the transpose flags, so we manually
    # transpose X and Y appropriately.
    if transpose_X:
        if X.ndim == 1:
            X = X.reshape((X.size, 1))
        elif X.ndim == 2:
            X = X.T
        else:
39
            dim = list(range(len(X.shape)))
40 41 42 43 44 45 46 47
            dim[-1], dim[len(X.shape) - 2] = dim[len(X.shape) - 2], dim[-1]
            X = np.transpose(X, tuple(dim))
    if transpose_Y:
        if Y.ndim == 1:
            Y = Y.reshape((1, Y.size))
        elif Y.ndim == 2:
            Y = Y.T
        else:
48
            dim = list(range(len(Y.shape)))
49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65
            dim[-1], dim[len(Y.shape) - 2] = dim[len(Y.shape) - 2], dim[-1]
            Y = np.transpose(Y, tuple(dim))

    if X.ndim == 3 and Y.ndim == 2:
        x_dims = X.shape
        X = X.reshape((x_dims[0] * x_dims[1], x_dims[2]))
    if Y.ndim == 3 and X.ndim == 2:
        y_dims = Y.shape
        Y = Y.reshape((y_dims[0] * y_dims[1], y_dims[2]))
    Out = np.matmul(X, Y)
    if not Out.shape:
        # We do not support 0-dimensional Tensors (scalars). So where
        # np.matmul outputs a scalar, we must convert to a Tensor of
        # shape (1, ) instead.
        # Everywhere else, we are compatible with np.matmul.
        Out = np.array([Out], dtype="float32")
    return Out
66

67

68 69 70
def generate_compatible_shapes(
    dim_X, dim_Y, transpose_X, transpose_Y, batch_size
):
71
    BATCH_SIZE = 2
72
    if batch_size is not None:
73 74
        BATCH_SIZE = batch_size

75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103
    M = 3
    N = 4
    K = 5
    if (dim_X == 1 and transpose_X) or (dim_Y == 1 and transpose_Y):
        K = 1
    if dim_X == 1:
        if transpose_X:
            shape_X = [M]
        else:
            shape_X = [K]
    if dim_Y == 1:
        if transpose_Y:
            shape_Y = [N]
        else:
            shape_Y = [K]
    if dim_X >= 2:
        if transpose_X:
            shape_X = [K, M]
        else:
            shape_X = [M, K]
    if dim_X == 3:
        shape_X = [BATCH_SIZE] + shape_X
    if dim_Y >= 2:
        if transpose_Y:
            shape_Y = [N, K]
        else:
            shape_Y = [K, N]
    if dim_Y == 3:
        shape_Y = [BATCH_SIZE] + shape_Y
104 105

    if dim_Y == 3 and dim_X == 2:
106
        if not transpose_X:
107 108 109 110
            shape_X[1] = shape_X[1] * BATCH_SIZE
        else:
            shape_X[0] = shape_X[0] * BATCH_SIZE

111 112 113
    return shape_X, shape_Y


114 115 116 117 118 119
def generate_compatible_shapes_2(dim, transpose_X, transpose_Y):
    M = 2
    N = 4
    K = 3
    shape_X = [2 for _ in range(dim - 2)]
    shape_Y = [2 for _ in range(dim - 2)]
120

121 122 123 124
    if transpose_X:
        shape_X += [K, M]
    else:
        shape_X += [M, K]
125

126 127 128 129
    if transpose_Y:
        shape_Y += [N, K]
    else:
        shape_Y += [K, N]
130

131
    return shape_X, shape_Y
132 133


134 135 136 137
class XPUTestMatmulOpErr(XPUOpTestWrapper):
    def __init__(self):
        self.op_name = "matmul"
        self.use_dynamic_create_class = False
138

139 140 141
    class API_TestMm(unittest.TestCase):
        def test_out(self):
            with fluid.program_guard(fluid.Program()):
142 143 144 145 146
                x = paddle.static.data(name="x", shape=[2], dtype=self.in_type)
                y = paddle.static.data(name='y', shape=[2], dtype=self.in_type)
                res = paddle.static.data(
                    name="output", shape=[1], dtype=self.in_type
                )
147 148 149 150
                result = paddle.mm(x, y)
                exe = fluid.Executor(fluid.XPUPlace(0))
                data1 = np.random.rand(2).astype(self.in_type)
                data2 = np.random.rand(2).astype(self.in_type)
151 152 153 154 155 156
                np_res = exe.run(
                    feed={'x': data1, 'y': data2}, fetch_list=[result]
                )
                expected_result = np.matmul(
                    data1.reshape(1, 2), data2.reshape(2, 1)
                )
157

158
                np.testing.assert_allclose(np_res, expected_result, atol=1e-3)
159 160 161 162 163 164 165 166 167 168

        def test_dygraph_without_out(self):
            device = fluid.XPUPlace(0)
            with fluid.dygraph.guard(device):
                input_array1 = np.random.rand(3, 4).astype(self.in_type)
                input_array2 = np.random.rand(4, 3).astype(self.in_type)
                data1 = fluid.dygraph.to_variable(input_array1)
                data2 = fluid.dygraph.to_variable(input_array2)
                out = paddle.mm(data1, data2)
                expected_result = np.matmul(input_array1, input_array2)
169 170 171
                np.testing.assert_allclose(
                    expected_result, out.numpy(), atol=1e-3
                )
172 173 174 175 176 177 178 179

    class Test_API_Matmul(unittest.TestCase):
        def test_dygraph_without_out(self):
            device = fluid.XPUPlace(0)
            with fluid.dygraph.guard(device):
                input_array1 = np.random.rand(3, 4).astype(self.in_type)
                input_array2 = np.random.rand(4, 3).astype(self.in_type)
                data1 = fluid.dygraph.to_variable(input_array1).astype(
180 181
                    self.in_type
                )
182
                data2 = fluid.dygraph.to_variable(input_array2).astype(
183 184
                    self.in_type
                )
185 186
                out = paddle.matmul(data1, data2)
                expected_result = np.matmul(input_array1, input_array2)
187 188 189
                np.testing.assert_allclose(
                    expected_result, out.numpy(), atol=1e-3
                )
190 191 192 193 194

    class API_TestMmError(unittest.TestCase):
        def test_errors(self):
            def test_error1():
                with fluid.program_guard(fluid.Program(), fluid.Program()):
195
                    data1 = paddle.static.data(
196 197
                        name="data1", shape=[10, 2], dtype="float32"
                    )
198
                    data2 = paddle.static.data(
199 200
                        name="data2", shape=[3, 10], dtype="float32"
                    )
201 202 203 204 205 206
                    paddle.mm(data1, data2)

            self.assertRaises(ValueError, test_error1)

            def test_error2():
                with fluid.program_guard(fluid.Program(), fluid.Program()):
207
                    data1 = paddle.static.data(
208 209
                        name="data1", shape=[-1, 10, 2], dtype="float32"
                    )
210
                    data2 = paddle.static.data(
211 212
                        name="data2", shape=[-1, 2, 10], dtype="float32"
                    )
213 214 215 216 217 218
                    paddle.mm(data1, data2)

            test_error2()

            def test_error3():
                with fluid.program_guard(fluid.Program(), fluid.Program()):
219
                    data1 = paddle.static.data(
220 221
                        name="data1", shape=[10, 10, 2], dtype="float32"
                    )
222
                    data2 = paddle.static.data(
223 224
                        name="data2", shape=[3, 2, 10], dtype="float32"
                    )
225 226 227 228 229 230 231 232
                    paddle.mm(data1, data2)

            self.assertRaises(ValueError, test_error3)


class TestMatmulBaseGenerator(XPUOpTest):
    def setUp(self):
        self.op_type = "matmul"
233 234 235
        self.dtype = (
            np.float32 if not hasattr(self, 'in_type') else self.in_type
        )
236

237 238 239 240 241
        self.__class__.no_need_check_grad = (
            False
            if not hasattr(self, 'no_need_check_grad')
            else self.no_need_check_grad
        )
242

243 244
        shape_X = [4, 5] if not hasattr(self, 'shape_X') else self.shape_X
        shape_Y = [5, 6] if not hasattr(self, 'shape_Y') else self.shape_Y
245 246 247 248 249 250
        transpose_X = (
            False if not hasattr(self, 'transpose_X') else self.transpose_X
        )
        transpose_Y = (
            False if not hasattr(self, 'transpose_Y') else self.transpose_Y
        )
251 252 253

        X = np.random.random(shape_X).astype(self.dtype)
        Y = np.random.random(shape_Y).astype(self.dtype)
254 255 256
        Out = reference_matmul(X, Y, transpose_X, transpose_Y).astype(
            self.dtype
        )
257 258 259 260 261 262 263 264 265
        self.inputs = {'X': X, 'Y': Y}
        self.attrs = {'transpose_X': transpose_X, 'transpose_Y': transpose_Y}
        self.outputs = {'Out': Out}

    def test_check_output(self):
        place = paddle.XPUPlace(0)
        self.check_output_with_place(place, atol=1e-3)

    def test_check_grad_normal(self):
266 267
        if (
            hasattr(self.__class__, "no_need_check_grad")
268
            and self.__class__.no_need_check_grad
269
        ):
270 271
            return

272
        place = paddle.XPUPlace(0)
273 274 275
        self.check_grad_with_place(
            place, ['X', 'Y'], 'Out', max_relative_error=5e-2
        )
276

277
    def test_check_grad_ignore_x(self):
278 279
        if (
            hasattr(self.__class__, "no_need_check_grad")
280
            and self.__class__.no_need_check_grad
281
        ):
282 283
            return

284
        place = paddle.XPUPlace(0)
285 286 287
        self.check_grad_with_place(
            place, ['Y'], 'Out', max_relative_error=5e-2, no_grad_set=set("X")
        )
288 289

    def test_check_grad_ignore_y(self):
290 291
        if (
            hasattr(self.__class__, "no_need_check_grad")
292
            and self.__class__.no_need_check_grad
293
        ):
294 295
            return

296
        place = paddle.XPUPlace(0)
297 298 299
        self.check_grad_with_place(
            place, ['X'], 'Out', max_relative_error=5e-2, no_grad_set=set('Y')
        )
300 301


302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317
class XPUTestMatmulOp1(XPUOpTestWrapper):
    def __init__(self):
        self.op_name = "matmul"
        self.use_dynamic_create_class = True

    def dynamic_create_class(self):
        base_class = TestMatmulBaseGenerator
        classes = []
        xpu_support_dims_list = [[1, 1], [2, 2], [3, 3]]
        batch_size = [2, 4, 5, 10, 50, 100, 300]
        for dims in xpu_support_dims_list:
            dim_X = dims[0]
            dim_Y = dims[1]
            for transose_x in [True, False]:
                for transose_y in [True, False]:
                    for batch in batch_size:
318 319 320
                        no_need_check_grad = False
                        if batch >= 5:
                            no_need_check_grad = True
321 322 323
                        class_name = 'TestMatMulOp_dimX_{}_dim_Y_{}_transX_{}_transY_{}_batch_{}'.format(
                            dim_X, dim_Y, transose_x, transose_y, batch
                        )
324
                        shape_x, shape_y = generate_compatible_shapes(
325 326
                            dim_X, dim_Y, transose_x, transose_y, batch
                        )
327 328 329 330 331
                        attr_dict = {
                            'shape_X': shape_x,
                            'shape_Y': shape_y,
                            'transpose_X': transose_x,
                            'transpose_Y': transose_y,
332
                            'no_need_check_grad': no_need_check_grad,
333
                            'op_type': "matmul",
334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350
                        }
                        classes.append([class_name, attr_dict])

        return base_class, classes


class XPUTestMatmulOp3(XPUOpTestWrapper):
    def __init__(self):
        self.op_name = "matmul"
        self.use_dynamic_create_class = True

    def dynamic_create_class(self):
        base_class = TestMatmulBaseGenerator
        classes = []
        for dim in [4]:
            for transpose_X in [False, True]:
                for transpose_Y in [False, True]:
351 352 353
                    class_name = 'TestMatMulOp2_dimX_{}_dim_Y_{}_transX_{}_transY_{}'.format(
                        dim, dim, transpose_X, transpose_Y
                    )
354
                    shape_X, shape_Y = generate_compatible_shapes_2(
355 356
                        dim, transpose_X, transpose_Y
                    )
357 358 359 360 361
                    attr_dict = {
                        'shape_X': shape_X,
                        'shape_Y': shape_Y,
                        'transpose_X': transpose_X,
                        'transpose_Y': transpose_Y,
362
                        'op_type': "matmul",
363 364 365 366 367 368 369 370 371 372
                    }
                    classes.append([class_name, attr_dict])
        return base_class, classes


support_types = get_xpu_op_support_types('matmul')
for stype in support_types:
    create_test_class(globals(), XPUTestMatmulOpErr, stype)
    create_test_class(globals(), XPUTestMatmulOp1, stype)
    create_test_class(globals(), XPUTestMatmulOp3, stype)
373 374

if __name__ == "__main__":
375
    paddle.enable_static()
376
    unittest.main()