diff --git a/paddle/fluid/operators/generator/cross_validate.py b/paddle/fluid/operators/generator/cross_validate.py index 82d21f47c62d0cba7e1593bac41c595fdc5a51bb..cfe013e73e4a2ec2f703c5fbf7aa32335c17816c 100644 --- a/paddle/fluid/operators/generator/cross_validate.py +++ b/paddle/fluid/operators/generator/cross_validate.py @@ -26,7 +26,7 @@ def main(forward_op_yaml_paths, backward_op_yaml_paths): with open(op_yaml_path, "rt", encoding="utf-8") as f: op_list = yaml.safe_load(f) if op_list is not None: - ops.update(to_named_dict((op_list))) + ops.update(to_named_dict(op_list)) cross_validate(ops) diff --git a/pyproject.toml b/pyproject.toml index 75bd21e7ba9aad26d43d70bd49116582f090626e..6519ad98c6f20a67a27befcc30bcce11ad788ef0 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -80,7 +80,7 @@ select = [ # "UP031", # "UP032", "UP033", - # "UP034", + "UP034", "UP035", "UP036", @@ -168,6 +168,11 @@ unfixable = [ ] [tool.ruff.per-file-ignores] +# Ignore unused imports in __init__.py "__init__.py" = ["F401"] +# Temporarily ignore test_slice.py to avoid PR-CI-CINN failure, please fix! +"python/paddle/fluid/tests/unittests/dygraph_to_static/test_slice.py" = ["UP034"] +# Ignore version check in setup.py "setup.py" = ["UP036"] +# Ignore unnecessary lambda in dy2st unittest test_lambda "python/paddle/fluid/tests/unittests/dygraph_to_static/test_lambda.py" = ["PLC3002"] diff --git a/python/paddle/distributed/auto_parallel/tuner/profiler.py b/python/paddle/distributed/auto_parallel/tuner/profiler.py index c419ae7fcf701ba58a1e2dc668c72b77ff6e95e4..138a8bddd41ab2c7af8e3fc6a37c1dfc7d2694d8 100644 --- a/python/paddle/distributed/auto_parallel/tuner/profiler.py +++ b/python/paddle/distributed/auto_parallel/tuner/profiler.py @@ -270,7 +270,7 @@ def profiler(args): with open(result_path, 'w') as fp: json.dump(result_dict, fp) - print("profile done! avg speed : {} step / s.".format((avg_tput))) + print("profile done! avg speed : {} step / s.".format(avg_tput)) except paddle.framework.core.EOFException: data_loader._inner_dataloader.reset() diff --git a/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py b/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py index 99a30333ac755dbafcf0795fc024ba5ba4ce66a2..21f3e26da741114257afc04760b6a55230c52be8 100644 --- a/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py +++ b/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py @@ -219,7 +219,7 @@ class AscendParserBase: tensor = core.GETensor(tensor_desc) data = ( - (value * np.ones((shape))) + (value * np.ones(shape)) .reshape(shape) .astype(self.ascend_helper.dtype2np(dtype)) ) @@ -282,7 +282,7 @@ class AscendParserBase: ) tensor = core.GETensor(tensor_desc) - data = np.ones((2)).astype("int32").reshape([2]) + data = np.ones(2).astype("int32").reshape([2]) data[0] = 64 buf = data.tobytes() data_8 = np.frombuffer(buf, dtype=np.uint8) diff --git a/python/paddle/distribution/normal.py b/python/paddle/distribution/normal.py index 22ebab0ed4b6205f0d5c0157f61a2e4a8fe5b1ea..edb08f6cfaeeeae4d2228f9ef639aac84d3092ea 100644 --- a/python/paddle/distribution/normal.py +++ b/python/paddle/distribution/normal.py @@ -241,7 +241,7 @@ class Normal(distribution.Distribution): ) return paddle.add( 0.5 + zero_tmp, - 0.5 * math.log(2 * math.pi) + paddle.log((self.scale + zero_tmp)), + 0.5 * math.log(2 * math.pi) + paddle.log(self.scale + zero_tmp), name=name, ) diff --git a/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py b/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py index 729ae93d6475d14e51468c2d0b5b92a3f688ec1b..4bc90d8f792e270f77a44f77437bb6aaa821210b 100644 --- a/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py +++ b/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py @@ -1185,7 +1185,7 @@ class TestStickBreakingTransform(unittest.TestCase): def test_codomain(self): self.assertTrue(isinstance(self._t._codomain, variable.Variable)) - @param.param_func(((np.random.random((10)),),)) + @param.param_func(((np.random.random(10),),)) def test_forward(self, input): np.testing.assert_allclose( self._t.inverse(self._t.forward(paddle.to_tensor(input))), @@ -1202,7 +1202,7 @@ class TestStickBreakingTransform(unittest.TestCase): def test_inverse_shape(self, shape, expected_shape): self.assertEqual(self._t.inverse_shape(shape), expected_shape) - @param.param_func(((np.random.random((10)),),)) + @param.param_func(((np.random.random(10),),)) def test_forward_log_det_jacobian(self, x): self.assertEqual( self._t.forward_log_det_jacobian(paddle.to_tensor(x)).shape, [1] diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py index c3bbd2b786526bb6c3507a4d94817036ba98c991..ea6cb541ff99ef8d26bc7d7dbc052c574c33d053 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py @@ -456,7 +456,7 @@ def _dygraph_fn(): x = np.random.random((1, 3)).astype('float32') with fluid.dygraph.guard(): fluid.dygraph.to_variable(x) - np.random.random((1)) + np.random.random(1) class TestDygraphApiRecognition(unittest.TestCase): diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py index e1d9185f57f9ce1e80c992361aaaa3edef1561d5..7a4340d805a5160f0bf30bc6e82e28731c2408fc 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py @@ -204,7 +204,7 @@ def test_optim_break_in_while(x): class TestContinueInFor(unittest.TestCase): def setUp(self): - self.input = np.zeros((1)).astype('int64') + self.input = np.zeros(1).astype('int64') self.place = ( fluid.CUDAPlace(0) if fluid.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py index ae7a081279946cec3ef7fdcf07f93be269af87dc..69100e60efb8de71bb4f66d88be4196f88238be3 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py @@ -138,7 +138,7 @@ class TestConvertWithCache(unittest.TestCase): @to_static def sum_even_until_limit(max_len, limit): - ret_sum = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) + ret_sum = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) for i in range(max_len): if i % 2 > 0: continue @@ -150,8 +150,8 @@ def sum_even_until_limit(max_len, limit): def sum_under_while(limit): - i = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) - ret_sum = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) + i = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) + ret_sum = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) while i <= limit: ret_sum += i i += 1 diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py index 0aeff4961fb4fd909d73307160a479e184ed6474..852b69e1a7b99b1dad28df715f284d26f95540d3 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py @@ -171,7 +171,7 @@ def test_dic_pop_2(x): class TestDictPop(unittest.TestCase): def setUp(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') self.place = ( paddle.CUDAPlace(0) if paddle.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py index cecfb00266af9b0dde320aef7e027bbcb60b9ba2..44e02950bc54819d90835aae85557a8267dfebed 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py @@ -219,7 +219,7 @@ class TestListWithoutControlFlow(unittest.TestCase): self.init_dygraph_func() def init_data(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') def init_dygraph_func(self): self.all_dygraph_funcs = [ @@ -275,7 +275,7 @@ class TestListInIf(TestListWithoutControlFlow): class TestListInWhileLoop(TestListWithoutControlFlow): def init_data(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') self.iter_num = 3 def init_dygraph_func(self): diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py index 0f72885fca937709c17c3b51512500e4f1cffb95..2796cc332df34cfedf6926a4c22713c6b4be66ce 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py @@ -265,7 +265,7 @@ def test_return_nested(x): class TestReturnBase(unittest.TestCase): def setUp(self): - self.input = np.ones((1)).astype('int32') + self.input = np.ones(1).astype('int32') self.place = ( fluid.CUDAPlace(0) if fluid.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py index 459b3719d2db2b83d0b4c3633a89890f2456cad3..47dadf6f71a87ad3d0d6ea275e253a4393dd446d 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py @@ -108,7 +108,7 @@ def train_static(args, batch_generator): # the best cross-entropy value with label smoothing loss_normalizer = -( (1.0 - args.label_smooth_eps) - * np.log((1.0 - args.label_smooth_eps)) + * np.log(1.0 - args.label_smooth_eps) + args.label_smooth_eps * np.log( args.label_smooth_eps / (args.trg_vocab_size - 1) + 1e-20 @@ -221,8 +221,7 @@ def train_dygraph(args, batch_generator): ) # the best cross-entropy value with label smoothing loss_normalizer = -( - (1.0 - args.label_smooth_eps) - * np.log((1.0 - args.label_smooth_eps)) + (1.0 - args.label_smooth_eps) * np.log(1.0 - args.label_smooth_eps) + args.label_smooth_eps * np.log(args.label_smooth_eps / (args.trg_vocab_size - 1) + 1e-20) ) diff --git a/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py b/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py index 46b03297cd1d2ae2d307513ebfc5cdf8e5147ebf..5e8489c1f5dd67f0f337c5fa295210b0e1f5b0aa 100644 --- a/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py +++ b/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py @@ -123,10 +123,8 @@ class TestBase(IPUOpTest): pad_batch = self.batch_size - dy_batch for k, v in feed.items(): pad_size = tuple( - ( - (0, 0 if i != 0 else pad_batch) - for i in range(len(v.shape)) - ) + (0, 0 if i != 0 else pad_batch) + for i in range(len(v.shape)) ) feed[k] = np.pad(v, pad_size, 'constant', constant_values=0) diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py b/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py index 6b30f83ae0c3120d13ce7e72cdc36643f7c642a7..397933737bac2c6096b3eeda2aabdb79bea533d9 100755 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py @@ -62,24 +62,20 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): if ( int( ( - ( - input_shape[2] - - (dilations[0] * (filter_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + input_shape[2] + - (dilations[0] * (filter_shape[2] - 1) + 1) ) + / strides[0] + + 1 ) <= 0 or int( ( - ( - input_shape[3] - - (dilations[1] * (filter_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + input_shape[3] + - (dilations[1] * (filter_shape[3] - 1) + 1) ) + / strides[1] + + 1 ) <= 0 ): @@ -88,28 +84,24 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): if ( int( ( - ( - input_shape[2] - + paddings[0] - + paddings[1] - - (dilations[0] * (filter_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + input_shape[2] + + paddings[0] + + paddings[1] + - (dilations[0] * (filter_shape[2] - 1) + 1) ) + / strides[0] + + 1 ) <= 0 or int( ( - ( - input_shape[3] - + paddings[2] - + paddings[3] - - (dilations[1] * (filter_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + input_shape[3] + + paddings[2] + + paddings[3] + - (dilations[1] * (filter_shape[3] - 1) + 1) ) + / strides[1] + + 1 ) <= 0 ): @@ -206,27 +198,23 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): f_shape[0], int( ( - ( - x_shape[2] - + padding[0] - + padding[1] - - (dilations[0] * (f_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + x_shape[2] + + padding[0] + + padding[1] + - (dilations[0] * (f_shape[2] - 1) + 1) ) + / strides[0] + + 1 ), int( ( - ( - x_shape[3] - + padding[2] - + padding[3] - - (dilations[1] * (f_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + x_shape[3] + + padding[2] + + padding[3] + - (dilations[1] * (f_shape[3] - 1) + 1) ) + / strides[1] + + 1 ), ] @@ -241,18 +229,14 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): x_shape[0], f_shape[0], int( - ( - (x_shape[2] - (dilations[0] * (f_shape[2] - 1) + 1)) - / strides[0] - + 1 - ) + (x_shape[2] - (dilations[0] * (f_shape[2] - 1) + 1)) + / strides[0] + + 1 ), int( - ( - (x_shape[3] - (dilations[1] * (f_shape[3] - 1) + 1)) - / strides[1] - + 1 - ) + (x_shape[3] - (dilations[1] * (f_shape[3] - 1) + 1)) + / strides[1] + + 1 ), ] bias_index = 1 diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py b/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py index 1ad4600df3ea5c8717daf5ff1f3cad356c9ded7a..cab24fb22178dad075bef2de5ac8e17389d9aca1 100644 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py @@ -42,7 +42,7 @@ class TestMkldnnPreluOp(MkldnnAutoScanTest): elif kwargs["mode"] == "channel": if len(kwargs['in_shape']) <= 1: # not valid case, just return 0 - return np.zeros((1)).astype(np.float32) + return np.zeros(1).astype(np.float32) if kwargs['data_format'] == 'NCHW': return np.random.random(kwargs['in_shape'][1]).astype( np.float32 @@ -54,7 +54,7 @@ class TestMkldnnPreluOp(MkldnnAutoScanTest): else: if len(kwargs['in_shape']) <= 1: # not valid case, just return 0 - return np.zeros((1)).astype(np.float32) + return np.zeros(1).astype(np.float32) return np.random.random(kwargs['in_shape']).astype(np.float32) prelu_op = OpConfig( diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py b/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py index 269a8edae19b305eb16351fdd8f477a8b59a7931..17c20e56de229c1a4451916fdeea1250374f72e8 100644 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py @@ -35,7 +35,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input1(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: @@ -45,7 +45,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input2(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: @@ -55,7 +55,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input3(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py index 1747184eaf9a687387156e05b1d9797dfc571e17..2d500a2e4fb244f1934ef8c979ad23959fa18133 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py @@ -90,7 +90,7 @@ class TestExpandV2ExpandShapesTensor1OneDNNOp(TestExpandV2OneDNNOp): self.expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): self.expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) def set_additional_inputs(self): diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py index ae68c3f5bbeb65b00d037edfc8d33a3964e61e67..a6852e3081943af08760ddb9f7313d90a70df034 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py @@ -64,7 +64,7 @@ class TestFCINT8OneDNNOp(OpTest): self.out_float = np.dot(self.x_float, self.y_float) if self.use_bias: - self.bias = np.random.random((10)).astype("float32") * 10 + self.bias = np.random.random(10).astype("float32") * 10 self.out_float += self.bias self.out_scale, self.out = self.quantize(self.out_float) diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py index fe1ca64dbe28a7554050f727ba0bcf92a9934890..50cf635b28a9f00f629fc0637ecd34b7a977b9b9 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py @@ -78,7 +78,7 @@ class TestFillZerosLike4DShapeTensorListPriorityOneDNNOp( self.shape_tensor_list = [] for index, elem in enumerate(shape): self.shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * elem) + ("x" + str(index), np.ones(1).astype('int32') * elem) ) self.inputs = {'ShapeTensorList': self.shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py index 07e7334fcd5ce7d7cb9eb3089c9f69ebe5f40cf0..07c4a5bf7ded23a6d83553c8da4d34a5de43fc0c 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py @@ -119,17 +119,17 @@ class TestDnnlMatMulOpMixedDimsXWiderTransposeX(TestDnnlMatMulWithGradOp): class TestDnnlMatMulOpVectorMultiply(TestDnnlMatMulWithGradOp): def generate_data(self): - self.x = np.random.random((5)).astype("float32") - self.y = np.random.random((5)).astype("float32") + self.x = np.random.random(5).astype("float32") + self.y = np.random.random(5).astype("float32") self.out = np.matmul(self.x, self.y) class TestDnnlMatMulOpVectorMultiplyTranspose(TestDnnlMatMulWithGradOp): def generate_data(self): - self.x = np.random.random((5)).astype("float32") + self.x = np.random.random(5).astype("float32") x_resized = np.copy(self.x) x_resized = np.expand_dims(x_resized, 1) - self.y = np.random.random((6)).astype("float32") + self.y = np.random.random(6).astype("float32") y_resized = np.copy(self.y) y_resized = np.expand_dims(y_resized, 0) self.out = np.matmul(x_resized, y_resized) diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py index c70fbd332043c3e257345c7a1ed854e97c05a9a1..13f1d0a7be76a01360c43e3c57ec543f4602eb78 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py @@ -103,14 +103,14 @@ class TestPReluModeChannelAlpha1DOneDNNOp(TestPReluModeChannelOneDNNOp): def init_attrs(self): self.mode = "channel" self.x = np.random.random((1, 100, 1)).astype("float32") - self.alpha = np.random.random((100)).astype("float32") + self.alpha = np.random.random(100).astype("float32") class TestPReluModeAllAlpha1DOneDNNOp(TestPReluModeAllOneDNNOp): def init_attrs(self): self.mode = "channel" self.x = np.random.random((1, 1, 100)).astype("float32") - self.alpha = np.random.random((1)).astype("float32") + self.alpha = np.random.random(1).astype("float32") # BF16 TESTS diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py index 7d16dec8583de329ada71f94fa0b04bd6a64e72e..8e607d5a4480f69ea7e2ea36110633aa8a2ef828 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py @@ -119,7 +119,7 @@ class TestReshape2OneDNNOpDimInfer1_attr_ShapeTensor(TestReshape2OneDNNOp): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs["ShapeTensor"] = shape_tensor @@ -138,7 +138,7 @@ class TestReshape2OneDNNOpDimInfer1_attr_ShapeTensorAndShape( shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs["Shape"] = np.array((1, 2, 3, 4), dtype="int32") diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py index 59d6a22a680ec3d5faa3f5991e91ebcebfdf1aea..b33ac0e34a0dd2d892ce0767d24f17b772417839 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py @@ -160,7 +160,7 @@ class TestSliceOneDNNOp_decs_dim_starts_ListTensor( def set_inputs(self): starts_tensor = [] for index, ele in enumerate(self.starts): - starts_tensor.append(("x1", np.ones((1)).astype('int32') * 2)) + starts_tensor.append(("x1", np.ones(1).astype('int32') * 2)) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} def config(self): diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py index 38292215563b2cfe4414bae66b481fb8808aa177..88fbdbde6a3a7ad11489760e62b1d2aa7ce4d43b 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py @@ -96,7 +96,7 @@ class TestSplitSectionsTensorBF16OneDNNOp(TestSplitSectionsBF16OneDNNOp): self.sections_tensor_list = [] for index, ele in enumerate(self.sections): self.sections_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.sections = [-1, -1, -1] indices_or_sections = [2, 3] # sections diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py index 88bedbb61cc83a6e903b8e259f092c04e18284ad..c1105b2096b6715249dcfd86fa3b6d2c1c3ef3bb 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py @@ -114,7 +114,7 @@ class TestSplitSectionsTensorOneDNNOp(TestSplitSectionsOneDNNOp): self.sections_tensor_list = [] for index, ele in enumerate(self.sections): self.sections_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.sections = [-1, -1, -1] indices_or_sections = [2, 3] # sections diff --git a/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py b/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py index 3dc34f8d099c9ea6f2699c113b76d6dc9724dfb9..64048753c3cacc6fc77821aedd656e31059ec9cb 100644 --- a/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py +++ b/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py @@ -70,7 +70,7 @@ class TestSequencePadOp(OpTest): start_idx = end_idx out_data = np.array(padded_sequences) - length = np.array(self.x_len_lod[0]).reshape((-1)) + length = np.array(self.x_len_lod[0]).reshape(-1) self.outputs = {'Out': out_data, 'Length': length} def setUp(self): @@ -90,7 +90,7 @@ class TestSequencePadOp2(TestSequencePadOp): def set_attr(self): self.x_shape = [12, 10] self.x_len_lod = [[2, 3, 4, 3]] - self.pad_value = np.random.random((10)) + self.pad_value = np.random.random(10) self.padded_length = -1 self.dtype = 'float64' @@ -108,7 +108,7 @@ class TestSequencePadOp4(TestSequencePadOp): def set_attr(self): self.x_shape = [12, 10] self.x_len_lod = [[2, 3, 4, 3]] - self.pad_value = np.random.random((10)) + self.pad_value = np.random.random(10) self.padded_length = 7 self.dtype = 'float64' diff --git a/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py b/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py index 930c25a2de495032f33313e16090a87b058cbbe5..daaa4b1ee9b3d0d270d6fc18395ccfbbe2a5c1ce 100644 --- a/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py +++ b/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py @@ -98,7 +98,7 @@ class TestSequenceUnpadOpError(unittest.TestCase): def test_length_variable(): x1 = paddle.static.data(name='x1', shape=[10, 5], dtype='float32') - len1 = np.random.random((10)).astype("int64") + len1 = np.random.random(10).astype("int64") paddle.static.nn.sequence_lod.sequence_pad(x=x1, length=len1) self.assertRaises(TypeError, test_length_variable) diff --git a/python/paddle/fluid/tests/unittests/test_activation_op.py b/python/paddle/fluid/tests/unittests/test_activation_op.py index 1badfb8e3f7722f8d060b6ed8512dca3cec24dd3..dfe0efabcc4a07d4391f53fbece7c1cd5a5ef886 100644 --- a/python/paddle/fluid/tests/unittests/test_activation_op.py +++ b/python/paddle/fluid/tests/unittests/test_activation_op.py @@ -2444,7 +2444,7 @@ class TestSoftRelu(TestActivation): t = np.copy(x) t[t < -threshold] = -threshold t[t > threshold] = threshold - out = np.log((np.exp(t) + 1)) + out = np.log(np.exp(t) + 1) self.inputs = {'X': OpTest.np_dtype_to_fluid_dtype(x)} self.attrs = {'threshold': threshold} diff --git a/python/paddle/fluid/tests/unittests/test_adamw_op.py b/python/paddle/fluid/tests/unittests/test_adamw_op.py index ea72c703f2304f316df6ccd0cfb61ecf2d8111c8..e94a0de7e41ffb86111a3faf3c9334c17d9bc92a 100644 --- a/python/paddle/fluid/tests/unittests/test_adamw_op.py +++ b/python/paddle/fluid/tests/unittests/test_adamw_op.py @@ -794,14 +794,14 @@ class TestAdamWOpLayerwiseLR(TestAdamWOp): out = linear1(x) out = linear2(out) - fc1_w_mon1 = np.zeros((linear1.weight.shape)).astype("float32") - fc1_w_mon2 = np.zeros((linear1.weight.shape)).astype("float32") - fc1_b_mon1 = np.zeros((linear1.bias.shape)).astype("float32") - fc1_b_mon2 = np.zeros((linear1.bias.shape)).astype("float32") - fc2_w_mon1 = np.zeros((linear2.weight.shape)).astype("float32") - fc2_w_mon2 = np.zeros((linear2.weight.shape)).astype("float32") - fc2_b_mon1 = np.zeros((linear2.bias.shape)).astype("float32") - fc2_b_mon2 = np.zeros((linear2.bias.shape)).astype("float32") + fc1_w_mon1 = np.zeros(linear1.weight.shape).astype("float32") + fc1_w_mon2 = np.zeros(linear1.weight.shape).astype("float32") + fc1_b_mon1 = np.zeros(linear1.bias.shape).astype("float32") + fc1_b_mon2 = np.zeros(linear1.bias.shape).astype("float32") + fc2_w_mon1 = np.zeros(linear2.weight.shape).astype("float32") + fc2_w_mon2 = np.zeros(linear2.weight.shape).astype("float32") + fc2_b_mon1 = np.zeros(linear2.bias.shape).astype("float32") + fc2_b_mon2 = np.zeros(linear2.bias.shape).astype("float32") cost = paddle.nn.functional.square_error_cost( input=out, label=y diff --git a/python/paddle/fluid/tests/unittests/test_addmm_op.py b/python/paddle/fluid/tests/unittests/test_addmm_op.py index 4e49a844d06834291065807713a2878968b4d2d6..3041841cdf84f165cb6479c3525a48568ff117f8 100644 --- a/python/paddle/fluid/tests/unittests/test_addmm_op.py +++ b/python/paddle/fluid/tests/unittests/test_addmm_op.py @@ -230,7 +230,7 @@ class TestAddMMOp4(OpTest): self.dtype = np.float64 self.init_dtype_type() self.inputs = { - 'Input': np.random.random((100)).astype(self.dtype), + 'Input': np.random.random(100).astype(self.dtype), 'X': np.random.random((20, 10)).astype(self.dtype), 'Y': np.random.random((10, 100)).astype(self.dtype), } @@ -296,7 +296,7 @@ class TestAddMMAPI(unittest.TestCase): self.assertRaises(ValueError, test_error1) def test_error2(): - data_x_wrong = np.ones((2)).astype(np.float32) + data_x_wrong = np.ones(2).astype(np.float32) x = paddle.to_tensor(data_x_wrong) y = paddle.to_tensor(data_y) input = paddle.to_tensor(data_input) @@ -318,7 +318,7 @@ class TestAddMMAPI(unittest.TestCase): self.assertRaises(ValueError, test_error3) def test_error4(): - data_input_wrong = np.ones((5)).astype(np.float32) + data_input_wrong = np.ones(5).astype(np.float32) x = paddle.to_tensor(data_x) y = paddle.to_tensor(data_y) input = paddle.to_tensor(data_input_wrong) @@ -358,7 +358,7 @@ class TestAddMMAPI(unittest.TestCase): def test_api_normal_2(self): data_x = np.ones((3, 10)).astype(np.float32) data_y = np.ones((10, 3)).astype(np.float32) - data_input = np.ones((3)).astype(np.float32) + data_input = np.ones(3).astype(np.float32) data_alpha = 0.1 data_beta = 1.0 @@ -383,7 +383,7 @@ class TestAddMMAPI(unittest.TestCase): def test_api_normal_3(self): data_x = np.ones((3, 10)).astype(np.float32) data_y = np.ones((10, 3)).astype(np.float32) - data_input = np.ones((1)).astype(np.float32) + data_input = np.ones(1).astype(np.float32) data_alpha = 0.1 data_beta = 1.0 diff --git a/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py b/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py index 11303d6e1a01b045d91e557865eaae3befb7bb3a..4926b10b4a1b268069e2700fbf197cbf6f34b96c 100644 --- a/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py +++ b/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py @@ -32,7 +32,7 @@ class TestCheckFiniteAndUnscaleOp(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] self.init_dtype() x = np.random.random((1024, 1024)).astype(self.dtype) - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -55,7 +55,7 @@ class TestCheckFiniteAndUnscaleOpWithNan(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] x = np.random.random((1024, 1024)).astype(self.dtype) x[128][128] = np.nan - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -80,7 +80,7 @@ class TestCheckFiniteAndUnscaleOpWithInf(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] x = np.random.random((1024, 1024)).astype(self.dtype) x[128][128] = np.inf - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { diff --git a/python/paddle/fluid/tests/unittests/test_bernoulli_op.py b/python/paddle/fluid/tests/unittests/test_bernoulli_op.py index 33a450310d1efda867317142d21f2d7113b444ff..af08b07237ff452c352eefb39dff0f684183d50a 100644 --- a/python/paddle/fluid/tests/unittests/test_bernoulli_op.py +++ b/python/paddle/fluid/tests/unittests/test_bernoulli_op.py @@ -24,7 +24,7 @@ def output_hist(out): hist, _ = np.histogram(out, bins=2) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.5 * np.ones((2)) + prob = 0.5 * np.ones(2) return hist, prob diff --git a/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py b/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py index f4b051e14519f21b9fd68d9949fa2eb61670159d..33104d3cf2d021884a6110d070eb8f8ea9c9d94e 100755 --- a/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py @@ -446,7 +446,7 @@ class TestBilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py index 1b3f615fde3d6795c2a5f332364d58e9f8fe39ec..b33e84bac872e2ea358931bfed0e25335d43f690 100755 --- a/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py @@ -798,7 +798,7 @@ class TestBilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_box_coder_op.py b/python/paddle/fluid/tests/unittests/test_box_coder_op.py index 889026aeb30bb3e15acd5d047bccc11a11ad516d..c0beed3ca4a9170ed7009781066bf70581552f70 100644 --- a/python/paddle/fluid/tests/unittests/test_box_coder_op.py +++ b/python/paddle/fluid/tests/unittests/test_box_coder_op.py @@ -293,7 +293,7 @@ class TestBoxCoderOpWithVariance(OpTest): self.python_api = wrapper_box_coder lod = [[1, 1, 1, 1, 1]] prior_box = np.random.random((30, 4)).astype('float32') - prior_box_var = np.random.random((4)).astype('float32') + prior_box_var = np.random.random(4).astype('float32') target_box = np.random.random((30, 81, 4)).astype('float32') code_type = "DecodeCenterSize" box_normalized = False @@ -325,7 +325,7 @@ class TestBoxCoderOpWithVarianceDygraphAPI(unittest.TestCase): def setUp(self): self.lod = [[1, 1, 1, 1, 1]] self.prior_box = np.random.random((30, 4)).astype('float32') - self.prior_box_var = np.random.random((4)).astype('float32') + self.prior_box_var = np.random.random(4).astype('float32') self.target_box = np.random.random((30, 81, 4)).astype('float32') self.code_type = "DecodeCenterSize" self.box_normalized = False diff --git a/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py b/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py index a75eb0f77a56da79f65c7252af67b5517a349b96..faf33532c33adb32dd497aecb298e12c0f0f2305 100644 --- a/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py +++ b/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py @@ -106,7 +106,7 @@ class TestAllocContinuousSpace(OpTest): coalesce_tensor_var = np.concatenate([input for input in inputs]) if set_constant: - coalesce_tensor_var = np.ones((len(coalesce_tensor_var))) * constant + coalesce_tensor_var = np.ones(len(coalesce_tensor_var)) * constant outputs = [ (out[0], np.ones(out[1].shape).astype(self.dtype) * constant) for out in outputs diff --git a/python/paddle/fluid/tests/unittests/test_compare_op.py b/python/paddle/fluid/tests/unittests/test_compare_op.py index 1ddc402c67b86761a18b46adb50a95b4414d45f1..d7dc124c5a33b8c6b0d1af5bbc4b738814e578ac 100755 --- a/python/paddle/fluid/tests/unittests/test_compare_op.py +++ b/python/paddle/fluid/tests/unittests/test_compare_op.py @@ -332,7 +332,7 @@ def create_paddle_case(op_type, callback): op = eval("paddle.%s" % (self.op_type)) out = op(x, y) exe = paddle.static.Executor(self.place) - input_x = np.arange(0, 5).reshape((5)).astype(np.int32) + input_x = np.arange(0, 5).reshape(5).astype(np.int32) input_y = np.array([5, 3, 2]).reshape((3, 1)).astype(np.int32) real_result = callback(input_x, input_y) (res,) = exe.run( diff --git a/python/paddle/fluid/tests/unittests/test_complex_matmul.py b/python/paddle/fluid/tests/unittests/test_complex_matmul.py index 65924da12b0ff00d3b1207505a4ff658109b6424..c6a89bb3cecbe42bbfa2af5cb5822500b61088ec 100644 --- a/python/paddle/fluid/tests/unittests/test_complex_matmul.py +++ b/python/paddle/fluid/tests/unittests/test_complex_matmul.py @@ -110,8 +110,8 @@ class TestComplexMatMulLayer(unittest.TestCase): x = np.random.random((2, 1, 100)).astype( dtype ) + 1j * np.random.random((2, 1, 100)).astype(dtype) - y = np.random.random((100)).astype(dtype) + 1j * np.random.random( - (100) + y = np.random.random(100).astype(dtype) + 1j * np.random.random( + 100 ).astype(dtype) np_result = np.matmul(x, y) diff --git a/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py b/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py index 31141142ccf4dfee32d9f4b16a52740b77b04eb9..b210ffb5d68fe27e13df7c3456004e0fa51d10eb 100644 --- a/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py +++ b/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py @@ -147,7 +147,7 @@ class TestCropTensorOpTensorAttr(OpTest): shape_tensor = [] for index, ele in enumerate(self.crop_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { 'X': np.random.random(self.x_shape).astype("float64"), @@ -159,7 +159,7 @@ class TestCropTensorOpTensorAttr(OpTest): offsets_tensor = [] for index, ele in enumerate(self.offsets): offsets_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { 'X': np.random.random(self.x_shape).astype("float64"), diff --git a/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py b/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py index 9bf2e65a6f48a26617a94b4be7a06bd3c799ca5a..6070f5622af511dcb2c2d3969f7b9f0f53c15d83 100644 --- a/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py +++ b/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py @@ -860,7 +860,7 @@ class CrossEntropyLoss(unittest.TestCase): N = 100 C = 200 input_np = np.random.random([N, C]).astype(self.dtype) - label_np = -np.ones((N)).astype(np.int64) + label_np = -np.ones(N).astype(np.int64) paddle.enable_static() prog = fluid.Program() startup_prog = fluid.Program() diff --git a/python/paddle/fluid/tests/unittests/test_cumsum_op.py b/python/paddle/fluid/tests/unittests/test_cumsum_op.py index a56b0f483bc084879db0f85bf88d0fa8e8423cc0..541387159cf038433ac1a40a608c79f064e8e588 100644 --- a/python/paddle/fluid/tests/unittests/test_cumsum_op.py +++ b/python/paddle/fluid/tests/unittests/test_cumsum_op.py @@ -186,7 +186,7 @@ class TestSumOp6(TestSumOp1): class TestSumOp7(TestSumOp1): def set_attrs_input_output(self): - self.x = np.random.random((100)).astype(self.dtype_) + self.x = np.random.random(100).astype(self.dtype_) self.out = self.x.cumsum(axis=0) diff --git a/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py b/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py index 0f0ecf3735dcc975b76d9513b7b531e04f92dc77..b443743a3515e3af1c3db1a998b88f370dafa8c9 100644 --- a/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py +++ b/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py @@ -73,7 +73,7 @@ class TestCumulativeTrapezoidWithOutDxX(TestCumulativeTrapezoidAPI): class TestCumulativeTrapezoidBroadcast(TestCumulativeTrapezoidAPI): def set_args(self): self.y = np.random.random((3, 3, 4)).astype('float32') - self.x = np.random.random((3)).astype('float32') + self.x = np.random.random(3).astype('float32') self.dx = None self.axis = 1 diff --git a/python/paddle/fluid/tests/unittests/test_dropout_op.py b/python/paddle/fluid/tests/unittests/test_dropout_op.py index 3644801f608745de90d9b91fae2b01bd7b8d2a57..6c5d88155bcd412eefa14771ce5d6c73a6e2122e 100644 --- a/python/paddle/fluid/tests/unittests/test_dropout_op.py +++ b/python/paddle/fluid/tests/unittests/test_dropout_op.py @@ -96,7 +96,7 @@ class TestDropoutOpInput1d(OpTest): self.attrs = {'dropout_prob': 0.0, 'fix_seed': True, 'is_test': False} self.outputs = { 'Out': self.inputs['X'], - 'Mask': np.ones((2000)).astype('uint8'), + 'Mask': np.ones(2000).astype('uint8'), } # Because prim op compare res with dygraph # when p = 0 dropout api return x,in dygraph mode x_grad = out_grad, diff --git a/python/paddle/fluid/tests/unittests/test_eig_op.py b/python/paddle/fluid/tests/unittests/test_eig_op.py index 5d3d556a67f0f121379ce210c8b1d9212f472960..c2861fc8bbebf94a24d85bb0a3ada43de2129e58 100644 --- a/python/paddle/fluid/tests/unittests/test_eig_op.py +++ b/python/paddle/fluid/tests/unittests/test_eig_op.py @@ -348,7 +348,7 @@ class TestEigWrongDimsError(unittest.TestCase): def test_error(self): paddle.device.set_device("cpu") paddle.disable_static() - a = np.random.random((3)).astype('float32') + a = np.random.random(3).astype('float32') x = paddle.to_tensor(a) self.assertRaises(ValueError, paddle.linalg.eig, x) diff --git a/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py b/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py index bf21a60521a07a59efe83bb9bddf0d6d9c700a42..2e518a2afb3893601e0bfbc4da21e0bdb55ec874 100644 --- a/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py +++ b/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py @@ -106,7 +106,7 @@ class TestElementWiseAddOp(unittest.TestCase): def test_check_forward_backward_with_scale_and_bias(self): np.random.seed(123) self.x = np.random.random((4, 32, 220, 220)).astype(np.float32) - self.y = np.random.random((32)).astype(np.float32) + self.y = np.random.random(32).astype(np.float32) self.out = self.x + self.y.reshape(1, 32, 1, 1) self.axis = 1 self.check_forward_backward() diff --git a/python/paddle/fluid/tests/unittests/test_empty_op.py b/python/paddle/fluid/tests/unittests/test_empty_op.py index 197568f830521bea33ec5a968ac31a34cbfaad5f..9f8c2a1eaebf3a70590ad1d4fc2a660a713138db 100644 --- a/python/paddle/fluid/tests/unittests/test_empty_op.py +++ b/python/paddle/fluid/tests/unittests/test_empty_op.py @@ -160,7 +160,7 @@ class TestEmptyOp_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/test_expand_op.py b/python/paddle/fluid/tests/unittests/test_expand_op.py index 420aeff5169812ab1c40f7f5756c7a47114a132e..f4d6258d82025cbf787d9234c083ebf41bce8f63 100644 --- a/python/paddle/fluid/tests/unittests/test_expand_op.py +++ b/python/paddle/fluid/tests/unittests/test_expand_op.py @@ -87,7 +87,7 @@ class TestExpandOpRank1_tensor_attr(OpTest): expand_times_tensor = [] for index, ele in enumerate(self.expand_times): expand_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_expand_v2_op.py b/python/paddle/fluid/tests/unittests/test_expand_v2_op.py index 3a17a4f1bc8d9f755703aebe247d5fb8b6a846be..5c0f6ff707fb45ee1f2f432c8635dd587ff982b9 100644 --- a/python/paddle/fluid/tests/unittests/test_expand_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_expand_v2_op.py @@ -89,7 +89,7 @@ class TestExpandV2OpRank1_tensor_attr(OpTest): expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_fc_op.py b/python/paddle/fluid/tests/unittests/test_fc_op.py index 47cf8c46a587e200417e93b1f6333e63c146873a..44b8e2f599d8ea469b66e722926246ff1fef4c34 100644 --- a/python/paddle/fluid/tests/unittests/test_fc_op.py +++ b/python/paddle/fluid/tests/unittests/test_fc_op.py @@ -51,7 +51,7 @@ class MatrixGenerate: if bias_dims == 2: self.bias = np.random.random((1, oc)).astype("float32") else: - self.bias = np.random.random((oc)).astype("float32") + self.bias = np.random.random(oc).astype("float32") class TestFCOp(OpTest): diff --git a/python/paddle/fluid/tests/unittests/test_fill_constant_op.py b/python/paddle/fluid/tests/unittests/test_fill_constant_op.py index aea337e67c1e83749f856edab2a5ac6951341e8e..8aeb64c1a46850a53295d8902f7ab0b059ad0e8c 100644 --- a/python/paddle/fluid/tests/unittests/test_fill_constant_op.py +++ b/python/paddle/fluid/tests/unittests/test_fill_constant_op.py @@ -155,7 +155,7 @@ class TestFillConstantOp1_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} @@ -180,7 +180,7 @@ class TestFillConstantOp2_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/test_full_like_op.py b/python/paddle/fluid/tests/unittests/test_full_like_op.py index c5717d69931f7082057d76633c9d32c5679d77d4..63bfb7be24fe0826b491e2f46b8c97ecf18262c6 100644 --- a/python/paddle/fluid/tests/unittests/test_full_like_op.py +++ b/python/paddle/fluid/tests/unittests/test_full_like_op.py @@ -68,7 +68,7 @@ class TestFullOp(unittest.TestCase): paddle.disable_static() input = paddle.arange(6, 10, dtype='float32') out = paddle.full_like(input, fill_value=888.88, dtype='float32') - out_numpy = np.random.random((4)).astype("float32") + out_numpy = np.random.random(4).astype("float32") out_numpy.fill(888.88) self.assertTrue((out.numpy() == out_numpy).all(), True) paddle.enable_static() @@ -77,7 +77,7 @@ class TestFullOp(unittest.TestCase): paddle.disable_static() input = paddle.arange(6, 10, dtype='float32') out = paddle.full_like(input, fill_value=float('inf')) - out_numpy = np.random.random((4)).astype("float32") + out_numpy = np.random.random(4).astype("float32") out_numpy.fill(float('inf')) self.assertTrue((out.numpy() == out_numpy).all(), True) paddle.enable_static() diff --git a/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py b/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py index aa49b29916e643c55a233ddc8b4538511368ebbc..c8739c38e7699d31b1aa034068909dd82f762a81 100644 --- a/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py +++ b/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py @@ -275,16 +275,16 @@ class APITestStaticFusedFFN(unittest.TestCase): linear1_weight_data = np.random.random( (d_model, dim_feedforward) ).astype(dtype) - linear1_bias_data = np.zeros((dim_feedforward)).astype(dtype) + linear1_bias_data = np.zeros(dim_feedforward).astype(dtype) linear2_weight_data = np.random.random( (dim_feedforward, d_model) ).astype(dtype) - linear2_bias_data = np.zeros((d_model)).astype(dtype) + linear2_bias_data = np.zeros(d_model).astype(dtype) - ln1_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln1_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) - ln2_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln2_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) + ln1_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln1_bias_data = np.zeros(d_model).astype(layer_norm_dtype) + ln2_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln2_bias_data = np.zeros(d_model).astype(layer_norm_dtype) res_list = [fused_out, ln_out] real_res = [] diff --git a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py index 95d74544e6730823fe242d3594166ab9548ae645..9830fe455b84ed40cadeb95fe7d9d5957dabe65b 100644 --- a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py +++ b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py @@ -100,7 +100,7 @@ class TestFusedGateAttentionOp(OpTest): self.gating_b = _random((self.num_heads, self.head_dim)) self.output_w = _random((self.num_heads, self.head_dim, self.out_dim)) - self.output_b = _random((self.out_dim)) + self.output_b = _random(self.out_dim) self.dout = _random( (self.batch_size, self.msa_len, self.res_len, self.q_dim) diff --git a/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py b/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py index 1cb9a9c917234b3b2dfeb8ad85da249508e683be..3b8bf23e6003bd061ceba2878fccb26e7c8716b8 100644 --- a/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py @@ -123,7 +123,7 @@ class TestGaussianRandomOp_ShapeTensorList(TestGaussianRandomOp): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.attrs = { diff --git a/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py b/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py index 2d7e6508661c0cc8b1233b45aa1257a7bb7fb9a6..98234e1613f597b01a207468a27a5062a0630791 100644 --- a/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py @@ -136,7 +136,7 @@ class TestGRUOp(OpTest): (self.num_layers * self.direction_num, batch_size, self.hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -162,7 +162,7 @@ class TestGRUOp(OpTest): self.outputs = { 'Out': output, 'State': [('last_hidden', last_hidden)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py b/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py index 5525854b00891ac47210d66959809bb91acf1e40..c0396a7a02929b9526e550b39d269ef45146c97a 100644 --- a/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py +++ b/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py @@ -186,7 +186,7 @@ class TestGumbelSoftmaxOpSampleDistribution(OpTest): # Construct statistics z for samples and # z is approximately N(0,1) for unbiased count expected = self.probs * self.shape[0] - z = (self.counts - expected) / np.sqrt((expected * (1 - self.probs))) + z = (self.counts - expected) / np.sqrt(expected * (1 - self.probs)) # A (lazy) approximate 99% two-sided test: # occurs with prob alpha~>=0.01 if unbiased self.assertLess(np.max(np.abs(z)).item(), 2.58) diff --git a/python/paddle/fluid/tests/unittests/test_imperative_framework.py b/python/paddle/fluid/tests/unittests/test_imperative_framework.py index 8f4f73daa0a3dd7f825663a96b3e108d675a6f20..3c81da2c6d59f33e14042918d61c3ca3d754cf62 100644 --- a/python/paddle/fluid/tests/unittests/test_imperative_framework.py +++ b/python/paddle/fluid/tests/unittests/test_imperative_framework.py @@ -64,7 +64,7 @@ class TestDygraphFramework(unittest.TestCase): "backward should not be usable in static graph mode" ) except AssertionError as e: - self.assertTrue((e is not None)) + self.assertTrue(e is not None) def test_dygraph_to_string(self): np_inp = np.array([[1.0, 2.0], [3.0, 4.0]], dtype=np.float32) diff --git a/python/paddle/fluid/tests/unittests/test_initializer.py b/python/paddle/fluid/tests/unittests/test_initializer.py index b87444e19dd4fd7ae06c4f557de9b5ac7105ee43..79b4782e6b6a301509809538df2408e5676c1b0f 100644 --- a/python/paddle/fluid/tests/unittests/test_initializer.py +++ b/python/paddle/fluid/tests/unittests/test_initializer.py @@ -38,7 +38,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-1, 1)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -632,7 +632,7 @@ class TestNumpyArrayInitializer(unittest.TestCase): program = framework.Program() block = program.global_block() - np_array = numpy.random.random((10000)).astype(dtype) + np_array = numpy.random.random(10000).astype(dtype) for _ in range(2): block.create_parameter( dtype=np_array.dtype, diff --git a/python/paddle/fluid/tests/unittests/test_initializer_nn.py b/python/paddle/fluid/tests/unittests/test_initializer_nn.py index b590fbe5212aa890099e75508975f6da0d137a1f..f90902631c532ba90439e062651bbb9298ff6c2a 100644 --- a/python/paddle/fluid/tests/unittests/test_initializer_nn.py +++ b/python/paddle/fluid/tests/unittests/test_initializer_nn.py @@ -644,7 +644,7 @@ class TestAssign(unittest.TestCase): program = framework.Program() block = program.global_block() - np_array = numpy.random.random((10000)).astype(dtype) + np_array = numpy.random.random(10000).astype(dtype) for _ in range(2): block.create_parameter( dtype=np_array.dtype, diff --git a/python/paddle/fluid/tests/unittests/test_jit_save_load.py b/python/paddle/fluid/tests/unittests/test_jit_save_load.py index e94c44193e7ca16149e893c0d2e2ca298c8760ac..d04357e201cbb4c11d3cbcde15f5eb69eeb678fa 100644 --- a/python/paddle/fluid/tests/unittests/test_jit_save_load.py +++ b/python/paddle/fluid/tests/unittests/test_jit_save_load.py @@ -1102,7 +1102,7 @@ class TestJitSaveLoadEmptyLayer(unittest.TestCase): def test_save_load_empty_layer(self): layer = EmptyLayer() - x = paddle.to_tensor(np.random.random((10)).astype('float32')) + x = paddle.to_tensor(np.random.random(10).astype('float32')) out = layer(x) paddle.jit.save(layer, self.model_path) load_layer = paddle.jit.load(self.model_path) @@ -1124,8 +1124,8 @@ class TestJitSaveLoadNoParamLayer(unittest.TestCase): def test_save_load_no_param_layer(self): layer = NoParamLayer() - x = paddle.to_tensor(np.random.random((5)).astype('float32')) - y = paddle.to_tensor(np.random.random((5)).astype('float32')) + x = paddle.to_tensor(np.random.random(5).astype('float32')) + y = paddle.to_tensor(np.random.random(5).astype('float32')) out = layer(x, y) paddle.jit.save(layer, self.model_path) load_layer = paddle.jit.load(self.model_path) @@ -1433,7 +1433,7 @@ class TestJitSaveLoadFinetuneLoad(unittest.TestCase): result_11 = layer_finetune(inps1) self.assertTrue(float((result_00 - result_10).abs().max()) < 1e-5) - self.assertTrue(float(((result_01 - result_11)).abs().max()) < 1e-5) + self.assertTrue(float((result_01 - result_11).abs().max()) < 1e-5) # NOTE(weixin): When there are multiple test functions in an diff --git a/python/paddle/fluid/tests/unittests/test_linear.py b/python/paddle/fluid/tests/unittests/test_linear.py index 4d4cdde6ede92f946830a0c66552717878b4bc25..8823a4f40be2e687f71f9651be2b12f59bb95d86 100644 --- a/python/paddle/fluid/tests/unittests/test_linear.py +++ b/python/paddle/fluid/tests/unittests/test_linear.py @@ -27,7 +27,7 @@ class LinearTestCase(unittest.TestCase): self.dtype = 'float32' self.input = np.ones((3, 1, 2)).astype(self.dtype) self.weight = np.ones((2, 2)).astype(self.dtype) - self.bias = np.ones((2)).astype(self.dtype) + self.bias = np.ones(2).astype(self.dtype) self.place = ( paddle.CUDAPlace(0) if core.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/test_linear_interp_op.py b/python/paddle/fluid/tests/unittests/test_linear_interp_op.py index db61b762c89e0a2e29bd481bdf8c05aadcc010c4..e18dd2057941b5a86f764f8f662417c59542f097 100755 --- a/python/paddle/fluid/tests/unittests/test_linear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_linear_interp_op.py @@ -227,7 +227,7 @@ class TestLinearInterpOpSizeTensor(TestLinearInterpOp): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py index 323e1d8ca5b924c0a50cd0c755bc17a28c370cbe..4bb0a210c6682883b122854efd308e6ae59d3ec5 100755 --- a/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py @@ -295,7 +295,7 @@ class TestLinearInterpOpSizeTensor(TestLinearInterpOp): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py b/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py index d1b944b29c336cda777cec00d57ee839630d4000..f81eb531424debfda9af45f1b9542f01525cc597 100644 --- a/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py @@ -474,7 +474,7 @@ class TestCUDNNLstmOp(OpTest): init_c = np.zeros((self.num_layers, batch_size, hidden_size)).astype( self.dtype ) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") if core.is_compiled_with_rocm(): for i in range(len(flat_w)): @@ -508,7 +508,7 @@ class TestCUDNNLstmOp(OpTest): 'Out': output, "LastH": last_hidden, 'LastC': last_cell, - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'StateOut': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_meshgrid_op.py b/python/paddle/fluid/tests/unittests/test_meshgrid_op.py index 2946008b0fd8b01429a0724a0ede2d1541d667c4..cc6f04eb7a971f44135d3c043e3f858dac7939b4 100644 --- a/python/paddle/fluid/tests/unittests/test_meshgrid_op.py +++ b/python/paddle/fluid/tests/unittests/test_meshgrid_op.py @@ -268,7 +268,7 @@ class TestMeshGrid_ZeroDim(TestMeshgridOp): self.shape = self.get_x_shape() ins = [] outs = [] - ins.append(np.random.random(([])).astype(self.dtype)) + ins.append(np.random.random([]).astype(self.dtype)) ins.append(np.random.random([2]).astype(self.dtype)) ins.append(np.random.random([3]).astype(self.dtype)) for i in range(len(self.shape)): diff --git a/python/paddle/fluid/tests/unittests/test_multi_dot_op.py b/python/paddle/fluid/tests/unittests/test_multi_dot_op.py index 0ac38b9ac08d2d7f05974660638799bb42cc3f06..a5555657c7f153f1b09f04119abedacd7271f191 100644 --- a/python/paddle/fluid/tests/unittests/test_multi_dot_op.py +++ b/python/paddle/fluid/tests/unittests/test_multi_dot_op.py @@ -104,7 +104,7 @@ class TestMultiDotOp4Mat(TestMultiDotOp): class TestMultiDotOpFirst1D(TestMultiDotOp): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -112,7 +112,7 @@ class TestMultiDotOpFirst1D(TestMultiDotOp): class TestMultiDotOp3MatFirst1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.C = np.random.random((3, 3)).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} @@ -121,7 +121,7 @@ class TestMultiDotOp3MatFirst1D(TestMultiDotOp3Mat): class TestMultiDotOp4MatFirst1D(TestMultiDotOp4Mat): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.C = np.random.random((3, 4)).astype(self.dtype) self.D = np.random.random((4, 5)).astype(self.dtype) @@ -139,7 +139,7 @@ class TestMultiDotOp4MatFirst1D(TestMultiDotOp4Mat): class TestMultiDotOpLast1D(TestMultiDotOp): def get_inputs_and_outputs(self): self.A = np.random.random((3, 6)).astype(self.dtype) - self.B = np.random.random((6)).astype(self.dtype) + self.B = np.random.random(6).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -148,7 +148,7 @@ class TestMultiDotOp3MatLast1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): self.A = np.random.random((2, 4)).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) - self.C = np.random.random((3)).astype(self.dtype) + self.C = np.random.random(3).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} self.outputs = {'Out': multi_dot([self.A, self.B, self.C])} @@ -163,7 +163,7 @@ class TestMultiDotOp4MatLast1D(TestMultiDotOp4Mat): self.A = np.random.random((2, 3)).astype(self.dtype) self.B = np.random.random((3, 2)).astype(self.dtype) self.C = np.random.random((2, 3)).astype(self.dtype) - self.D = np.random.random((3)).astype(self.dtype) + self.D = np.random.random(3).astype(self.dtype) self.inputs = { 'X': [ ('x0', self.A), @@ -178,7 +178,7 @@ class TestMultiDotOp4MatLast1D(TestMultiDotOp4Mat): class TestMultiDotOpFirstAndLast1D(TestMultiDotOp): def get_inputs_and_outputs(self): self.A = np.random.random((4,)).astype(self.dtype) - self.B = np.random.random((4)).astype(self.dtype) + self.B = np.random.random(4).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -187,7 +187,7 @@ class TestMultiDotOp3MatFirstAndLast1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): self.A = np.random.random((6,)).astype(self.dtype) self.B = np.random.random((6, 4)).astype(self.dtype) - self.C = np.random.random((4)).astype(self.dtype) + self.C = np.random.random(4).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} self.outputs = {'Out': multi_dot([self.A, self.B, self.C])} @@ -197,7 +197,7 @@ class TestMultiDotOp4MatFirstAndLast1D(TestMultiDotOp4Mat): self.A = np.random.random((3,)).astype(self.dtype) self.B = np.random.random((3, 4)).astype(self.dtype) self.C = np.random.random((4, 2)).astype(self.dtype) - self.D = np.random.random((2)).astype(self.dtype) + self.D = np.random.random(2).astype(self.dtype) self.inputs = { 'X': [ ('x0', self.A), diff --git a/python/paddle/fluid/tests/unittests/test_mv_op.py b/python/paddle/fluid/tests/unittests/test_mv_op.py index 93b541e7eccf0d6ae9e4094268f727a8892989e8..9f0840fb123831672cf3e7532ccc1fd669ce6b04 100644 --- a/python/paddle/fluid/tests/unittests/test_mv_op.py +++ b/python/paddle/fluid/tests/unittests/test_mv_op.py @@ -37,7 +37,7 @@ class TestMVOp(OpTest): def init_config(self): self.x = np.random.random((2, 100)).astype("float64") - self.vec = np.random.random((100)).astype("float64") + self.vec = np.random.random(100).astype("float64") class TestMVAPI(unittest.TestCase): @@ -46,7 +46,7 @@ class TestMVAPI(unittest.TestCase): self.x_data = np.random.random((5, 100)).astype("float64") self.x = paddle.to_tensor(self.x_data) - self.vec_data = np.random.random((100)).astype("float64") + self.vec_data = np.random.random(100).astype("float64") self.vec = paddle.to_tensor(self.vec_data) z = paddle.mv(self.x, self.vec) np_z = z.numpy() diff --git a/python/paddle/fluid/tests/unittests/test_nanmedian.py b/python/paddle/fluid/tests/unittests/test_nanmedian.py index 42509cdb5f2654d9609cf5f40833d6d4c14133e5..88950271d32b92867ad7a0e23533f193867a45c6 100644 --- a/python/paddle/fluid/tests/unittests/test_nanmedian.py +++ b/python/paddle/fluid/tests/unittests/test_nanmedian.py @@ -189,7 +189,7 @@ class TestNanmedian(unittest.TestCase): x_np[2, 3:] = np.nan x_np_sorted = np.sort(x_np) nan_counts = np.count_nonzero(np.isnan(x_np).astype(np.int32), axis=1) - np_grad = np.zeros((shape)) + np_grad = np.zeros(shape) for i in range(shape[0]): valid_cnts = shape[1] - nan_counts[i] if valid_cnts == 0: diff --git a/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py b/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py index 86233f7411ac10ed082e3357a9b8e0871c20afda..cb153bdc55b8e6ea4c6e905112883eb2f8f0aae4 100755 --- a/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py @@ -374,7 +374,7 @@ class TestNearestInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py index 2447818c896d32953338c886b63e4800f8262113..6ddbdea0ece07782e8fef4acddee6bdede718ecd 100755 --- a/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py @@ -874,7 +874,7 @@ class TestNearestInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_npair_loss_op.py b/python/paddle/fluid/tests/unittests/test_npair_loss_op.py index a4a1b16efc6d396797d0a85a2ebe03ea85ee023f..b49cdf0547c47eed365a81d9dcbee77fa3ea1881 100755 --- a/python/paddle/fluid/tests/unittests/test_npair_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_npair_loss_op.py @@ -128,7 +128,7 @@ class TestNpairLossOpError(unittest.TestCase): with program_guard(Program(), Program()): anchor_np = np.random.random((2, 4)).astype("float32") positive_np = np.random.random((2, 4)).astype("float32") - labels_np = np.random.random((2)).astype("float32") + labels_np = np.random.random(2).astype("float32") anchor_data = paddle.static.data( name='anchor', shape=[2, 4], dtype='float32' ) diff --git a/python/paddle/fluid/tests/unittests/test_numel_op.py b/python/paddle/fluid/tests/unittests/test_numel_op.py index e4122bc0fb1b7d63ba998631d48374e480a24784..81ddbdc45a02fffcf294aef011f0f202f8e05a14 100644 --- a/python/paddle/fluid/tests/unittests/test_numel_op.py +++ b/python/paddle/fluid/tests/unittests/test_numel_op.py @@ -27,7 +27,7 @@ class TestNumelOp(OpTest): self.op_type = "size" self.python_api = paddle.numel self.init() - x = np.random.random((self.shape)).astype(self.dtype) + x = np.random.random(self.shape).astype(self.dtype) self.inputs = { 'Input': x, } @@ -82,7 +82,7 @@ class TestNumelOpBF16(OpTest): self.python_api = paddle.numel self.dtype = np.uint16 self.init() - x = np.random.random((self.shape)).astype(np.float32) + x = np.random.random(self.shape).astype(np.float32) self.inputs = {'Input': convert_float_to_uint16(x)} self.outputs = {'Out': np.array([np.size(x)])} diff --git a/python/paddle/fluid/tests/unittests/test_pool2d_op.py b/python/paddle/fluid/tests/unittests/test_pool2d_op.py index f4cf8943ea3bf10fd2999167f35193f218477824..7a835f0fb440eabe58025c1dc46662fa00509d91 100644 --- a/python/paddle/fluid/tests/unittests/test_pool2d_op.py +++ b/python/paddle/fluid/tests/unittests/test_pool2d_op.py @@ -778,7 +778,7 @@ class TestCase4_AsyPadding(TestCase4): self.shape = [2, 3, 7, 7] -class TestCase5_AsyPadding((TestCase5)): +class TestCase5_AsyPadding(TestCase5): def init_test_case(self): self.ksize = [3, 3] self.strides = [1, 1] diff --git a/python/paddle/fluid/tests/unittests/test_pylayer_op.py b/python/paddle/fluid/tests/unittests/test_pylayer_op.py index b270a8b0d7f702ce38961d12f3e123d70f28026e..ce036905138ae73ac8101eeffe2d0aa3419213ab 100644 --- a/python/paddle/fluid/tests/unittests/test_pylayer_op.py +++ b/python/paddle/fluid/tests/unittests/test_pylayer_op.py @@ -55,7 +55,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_return_none_with_no_grad(self): @@ -91,7 +91,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_single_output(self): @@ -119,7 +119,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_multi_output(self): @@ -149,7 +149,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_pylayer_num_output_match(self): diff --git a/python/paddle/fluid/tests/unittests/test_randint_op.py b/python/paddle/fluid/tests/unittests/test_randint_op.py index 678cb3a5a2e649695c7217364697d06fc1f9aff5..073b8255be95793d9ab006e331d76e9bf745da64 100644 --- a/python/paddle/fluid/tests/unittests/test_randint_op.py +++ b/python/paddle/fluid/tests/unittests/test_randint_op.py @@ -29,7 +29,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-10, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -74,7 +74,7 @@ class TestRandintOp_attr_tensorlist(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -160,7 +160,7 @@ class TestRandintImperative(unittest.TestCase): x3 = paddle.tensor.random.randint(n) for i in [x1, x2, x3]: for j in i.numpy().tolist(): - self.assertTrue((j >= 0 and j < n)) + self.assertTrue(j >= 0 and j < n) paddle.enable_static() diff --git a/python/paddle/fluid/tests/unittests/test_randperm_op.py b/python/paddle/fluid/tests/unittests/test_randperm_op.py index b482a0666aaed9baf4f854c42a5e0f93866119ad..eaecf087f9f04b8fd25f3e363515f4c6d4f4dabf 100644 --- a/python/paddle/fluid/tests/unittests/test_randperm_op.py +++ b/python/paddle/fluid/tests/unittests/test_randperm_op.py @@ -63,7 +63,7 @@ class TestRandpermOp(OpTest): self.dtype = "int64" self.inputs = {} - self.outputs = {"Out": np.zeros((self.n)).astype(self.dtype)} + self.outputs = {"Out": np.zeros(self.n).astype(self.dtype)} self.init_attrs() self.attrs = { "n": self.n, diff --git a/python/paddle/fluid/tests/unittests/test_rank_attention_op.py b/python/paddle/fluid/tests/unittests/test_rank_attention_op.py index 8428e6a043e392a0bb76f2851e7980dbf830ac04..d699a6311a3f4dce588e234025d3a0e75f5aa7c8 100644 --- a/python/paddle/fluid/tests/unittests/test_rank_attention_op.py +++ b/python/paddle/fluid/tests/unittests/test_rank_attention_op.py @@ -24,7 +24,7 @@ from paddle.fluid import core def gen_input_help(input, rank_offset, max_rank, max_size): input_row, input_col = input.shape max_ins = np.max((max_size, input_row)) - input_help = np.zeros((max_ins * max_rank * input_col)) + input_help = np.zeros(max_ins * max_rank * input_col) ins_rank = np.zeros((max_ins, 1)) ins_rank.fill(-1) diff --git a/python/paddle/fluid/tests/unittests/test_reshape_op.py b/python/paddle/fluid/tests/unittests/test_reshape_op.py index 0647cde61cc21a7bc0514e41e0b8d4689ab70996..24b8661694edbfde50458dd36ef22b9ba9a3fef7 100755 --- a/python/paddle/fluid/tests/unittests/test_reshape_op.py +++ b/python/paddle/fluid/tests/unittests/test_reshape_op.py @@ -174,7 +174,7 @@ class TestReshapeOp_attr_ShapeTensor(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py b/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py index bf01589c577d3e1a19dc027c959d3c570b3e945c..a7104e373999a9d647d60d40c6831ea7993f2f74 100644 --- a/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py +++ b/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py @@ -120,7 +120,7 @@ class TestRnnError(unittest.TestCase): self.assertRaises(TypeError, test_initial_states_type) def test_sequence_length_type(): - np_sequence_length = np.random.random((batch_size)).astype( + np_sequence_length = np.random.random(batch_size).astype( "float32" ) dynamic_rnn( diff --git a/python/paddle/fluid/tests/unittests/test_rnn_op.py b/python/paddle/fluid/tests/unittests/test_rnn_op.py index f4fc13996fd28351119f9b0147f8e867f59c35cb..73ef33816534d932515991d50fc9bf7dab334761 100644 --- a/python/paddle/fluid/tests/unittests/test_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_rnn_op.py @@ -137,7 +137,7 @@ class TestRNNOp(OpTest): init_c = np.zeros( (self.num_layers * self.direction_num, batch_size, hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -163,7 +163,7 @@ class TestRNNOp(OpTest): self.outputs = { 'Out': output, "State": [('last_hidden', last_hidden), ('last_cell', last_cell)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_seed_op.py b/python/paddle/fluid/tests/unittests/test_seed_op.py index d72bc8601d7c499592545989bbe356ca405b8751..426dab42148132daae224c52f94ecdd24223358e 100644 --- a/python/paddle/fluid/tests/unittests/test_seed_op.py +++ b/python/paddle/fluid/tests/unittests/test_seed_op.py @@ -28,7 +28,7 @@ class TestSeedOpFixSeed(OpTest): self.op_type = "seed" self.inputs = {} self.attrs = {"seed": 123} - self.outputs = {"Out": np.asarray((123)).astype('int')} + self.outputs = {"Out": np.asarray(123).astype('int')} def test_check_output(self): self.check_output() @@ -39,7 +39,7 @@ class TestSeedOpDiffSeed(OpTest): self.op_type = "seed" self.inputs = {} self.attrs = {"seed": 0} - self.outputs = {"Out": np.asarray((123)).astype('int')} + self.outputs = {"Out": np.asarray(123).astype('int')} def test_check_output(self): self.check_output(no_check_set=["Out"]) diff --git a/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py b/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py index dfeeef8ad29579279e63b15e01ae67e74af55c2e..abf4ba677ba86da0f1e13905e9986028e0192dbd 100644 --- a/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py @@ -36,8 +36,8 @@ def sigmoid_focal_loss_forward( a = int(idx / num_classes) d = int(idx % num_classes) label = label_data[a] - c_pos = float((int(label) == int(d + 1))) - c_neg = float(((int(label) != -1) & (int(label) != (d + 1)))) + c_pos = float(int(label) == int(d + 1)) + c_neg = float((int(label) != -1) & (int(label) != (d + 1))) fg_num = max(fg_num_data, 1) z_neg = (1.0 - alpha) / fg_num z_pos = alpha / fg_num diff --git a/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py b/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py index 4fe7f5a9dd0078e79b33465aea82cb41619d9273..b0901005204500494cd22e12a2880b4883cbc320 100644 --- a/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py @@ -128,7 +128,7 @@ class TestSimpleRNNOp(OpTest): (self.num_layers * self.direction_num, batch_size, hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -154,7 +154,7 @@ class TestSimpleRNNOp(OpTest): self.outputs = { 'Out': output, 'State': [('last_hidden', last_hidden)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_slice_op.py b/python/paddle/fluid/tests/unittests/test_slice_op.py index 4e8e62b1f441e69990b27439c1b771f8d07c084b..0314a37170d0e1c7c11e64e8b726a34e66e4a3b2 100644 --- a/python/paddle/fluid/tests/unittests/test_slice_op.py +++ b/python/paddle/fluid/tests/unittests/test_slice_op.py @@ -172,7 +172,7 @@ class TestSliceOp_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int64') * ele) + ("x" + str(index), np.ones(1).astype('int64') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -212,7 +212,7 @@ class TestSliceOp_decs_dim_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -379,7 +379,7 @@ class TestSliceOp_starts_OneTensor_ends_ListTensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("y" + str(index), np.ones((1)).astype('int32') * ele) + ("y" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_split_op.py b/python/paddle/fluid/tests/unittests/test_split_op.py index 7d50d9586b3bf9eeb49b877fa30a7bd1a7845600..f0929e5758d8fb8f48e1adbcd7003b25df5705d7 100644 --- a/python/paddle/fluid/tests/unittests/test_split_op.py +++ b/python/paddle/fluid/tests/unittests/test_split_op.py @@ -154,7 +154,7 @@ class TestSplitOp_SectionsTensor(OpTest): sections_tensor = [] for index, ele in enumerate(self.sections): sections_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SectionsTensorList'] = sections_tensor diff --git a/python/paddle/fluid/tests/unittests/test_strided_slice_op.py b/python/paddle/fluid/tests/unittests/test_strided_slice_op.py index fb841244da14754f5fbe018132bdcf432b7bc9e9..dc5397fb4f4891cfc9f5e9440ac59d777e98ed67 100644 --- a/python/paddle/fluid/tests/unittests/test_strided_slice_op.py +++ b/python/paddle/fluid/tests/unittests/test_strided_slice_op.py @@ -324,7 +324,7 @@ class TestStridedSliceOp_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -366,7 +366,7 @@ class TestStridedSliceOp_ends_ListTensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'EndsTensorList': ends_tensor} @@ -477,7 +477,7 @@ class TestStridedSliceOp_listTensor_Tensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.op_type = "strided_slice" self.python_api = paddle.strided_slice diff --git a/python/paddle/fluid/tests/unittests/test_tile_op.py b/python/paddle/fluid/tests/unittests/test_tile_op.py index ec450cd888efdace43ec013bb0012dead0d8c1c0..61901ce1df4efd63de95b018ce2b4a5127c5f738 100644 --- a/python/paddle/fluid/tests/unittests/test_tile_op.py +++ b/python/paddle/fluid/tests/unittests/test_tile_op.py @@ -111,7 +111,7 @@ class TestTileOpRank1_tensor_attr(OpTest): repeat_times_tensor = [] for index, ele in enumerate(self.repeat_times): repeat_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_trapezoid.py b/python/paddle/fluid/tests/unittests/test_trapezoid.py index ed5bbc2ae21df1a900f1ac1ed984913170c58d44..f622c6b746663e9949c854f6d219e70475d88b75 100644 --- a/python/paddle/fluid/tests/unittests/test_trapezoid.py +++ b/python/paddle/fluid/tests/unittests/test_trapezoid.py @@ -148,7 +148,7 @@ class TestTrapezoidWithOutDxX(TestTrapezoidAPI): class TestTrapezoidBroadcast(TestTrapezoidAPI): def set_args(self): self.y = np.random.random((3, 3, 4)).astype('float32') - self.x = np.random.random((3)).astype('float32') + self.x = np.random.random(3).astype('float32') self.dx = None self.axis = 1 diff --git a/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py b/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py index 8001997696e46dcb43f2f45eb35809e94205e19a..031639fc080aabca351959196f7fd91cf61d3ff6 100755 --- a/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py @@ -528,7 +528,7 @@ class TestTrilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py index be13d5ebda1afda51f55b62b9581741c439d4f3e..6b774fafe3fb45ee8c60a69678d01524c85ba0c7 100755 --- a/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py @@ -888,7 +888,7 @@ class TestTrilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py b/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py index 0e08c3d8bc07f0e3136dc9ba8494ad50399f6ef3..e76bbc80033fa4fd406bb9592b831e87955f9959 100644 --- a/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py +++ b/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py @@ -70,7 +70,7 @@ class TestUniformRandomOpBF16AttrTensorList(TestUniformRandomOpBF16): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() diff --git a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py index 3f550e3ed7cd22191ca8eeac1a4a91344137ce67..de1984098b93848562fef764bbb6aa3942c1a2d1 100644 --- a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py @@ -33,7 +33,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-5, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -46,7 +46,7 @@ def output_hist_diag(out): hist, _ = np.histogram(out, range=(-5, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -58,7 +58,7 @@ class TestUniformRandomOp_attr_tensorlist(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -90,7 +90,7 @@ class TestUniformRandomOp_attr_tensorlist_int32(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int32") * ele) + ("x" + str(index), np.ones(1).astype("int32") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -471,7 +471,7 @@ class TestUniformRandomDygraphMode(unittest.TestCase): x = paddle.uniform([10], dtype="float32", min=0.0, max=1.0) x_np = x.numpy() for i in range(10): - self.assertTrue((x_np[i] > 0 and x_np[i] < 1.0)) + self.assertTrue(x_np[i] > 0 and x_np[i] < 1.0) class TestUniformRandomBatchSizeLikeOpError(unittest.TestCase): @@ -562,7 +562,7 @@ class TestUniformDygraphMode(unittest.TestCase): ) x_np = x.numpy() for i in range(10): - self.assertTrue((x_np[i] > 0 and x_np[i] < 1.0)) + self.assertTrue(x_np[i] > 0 and x_np[i] < 1.0) class TestUniformDtype(unittest.TestCase): diff --git a/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py b/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py index ea5044d1f81ed2f7e2fb7ccbad8132a153e2edcb..b7b4c185e97d980bbae8d8d1528b2d20666c574b 100755 --- a/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py +++ b/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py @@ -120,7 +120,7 @@ class TestUnsqueezeOp_AxesTensorList(OpTest): axes_tensor_list = [] for index, ele in enumerate(self.axes): axes_tensor_list.append( - ("axes" + str(index), np.ones((1)).astype('int32') * ele) + ("axes" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_variable.py b/python/paddle/fluid/tests/unittests/test_variable.py index 8e5752d1985ce291c497b04edadffdccd6950ba8..c26053a734297d037fe244c4ce709ab6f66835cf 100644 --- a/python/paddle/fluid/tests/unittests/test_variable.py +++ b/python/paddle/fluid/tests/unittests/test_variable.py @@ -173,7 +173,7 @@ class TestVariable(unittest.TestCase): y_1 = y[:, 0] feeder = fluid.DataFeeder(place=place, feed_list=[x]) data = [] - data.append((np.random.randint(10, size=[13]).astype('float32'))) + data.append(np.random.randint(10, size=[13]).astype('float32')) exe.run(fluid.default_startup_program()) local_out = exe.run( diff --git a/python/paddle/fluid/tests/unittests/test_yolo_box_op.py b/python/paddle/fluid/tests/unittests/test_yolo_box_op.py index a572b91b0b5bf9df82684b2eedadd051304195ae..81445c081fea11426456325a5e47fb6ee154292a 100644 --- a/python/paddle/fluid/tests/unittests/test_yolo_box_op.py +++ b/python/paddle/fluid/tests/unittests/test_yolo_box_op.py @@ -21,13 +21,13 @@ import paddle def sigmoid(x): - return 1.0 / (1.0 + np.exp(((-1.0) * x))) + return 1.0 / (1.0 + np.exp((-1.0) * x)) def YoloBox(x, img_size, attrs): (n, c, h, w) = x.shape anchors = attrs['anchors'] - an_num = int((len(anchors) // 2)) + an_num = int(len(anchors) // 2) class_num = attrs['class_num'] conf_thresh = attrs['conf_thresh'] downsample = attrs['downsample_ratio'] @@ -145,7 +145,7 @@ class TestYoloBoxOp(OpTest): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -166,7 +166,7 @@ class TestYoloBoxOp(OpTest): class TestYoloBoxOpNoClipBbox(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -187,7 +187,7 @@ class TestYoloBoxOpNoClipBbox(TestYoloBoxOp): class TestYoloBoxOpScaleXY(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -208,7 +208,7 @@ class TestYoloBoxOpScaleXY(TestYoloBoxOp): class TestYoloBoxOpIoUAware(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -295,7 +295,7 @@ class TestYoloBoxStatic(unittest.TestCase): class TestYoloBoxOpHW(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 diff --git a/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py b/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py index 288ed525a7831e1b3699d0b0fdea3ed29adb9739..79994d87e4266c1eb333bd0e091c1e18a7b36024 100644 --- a/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py @@ -81,7 +81,7 @@ def YOLOv3Loss(x, gtbox, gtlabel, gtscore, attrs): bias_x_y = -0.5 * (scale_x_y - 1.0) input_size = downsample_ratio * h x = x.reshape((n, mask_num, 5 + class_num, h, w)).transpose((0, 1, 3, 4, 2)) - loss = np.zeros((n)).astype('float64') + loss = np.zeros(n).astype('float64') smooth_weight = min(1.0 / class_num, 1.0 / 40) label_pos = 1.0 - smooth_weight if use_label_smooth else 1.0 diff --git a/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py index 9645fbb7d003331362771c16dacfb730bfc2ffdb..b071db95b40bc9ca71e184e95cc12e94ae09f64c 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py @@ -1011,7 +1011,7 @@ class XPUTestSoftReluOP(XPUOpTestWrapper): t = np.copy(x) t[t < -threshold] = -threshold t[t > threshold] = threshold - out = np.log((np.exp(t) + 1)) + out = np.log(np.exp(t) + 1) self.inputs = {'X': x} self.outputs = {'Out': out} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py index 0899d6f45e1b658de3f4602bd6903da1fa74cedc..d30fdbed09db8711e499fb9e19c32494b83bfff3 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py @@ -484,30 +484,22 @@ class XPUTestAdamwOp2(XPUOpTestWrapper): out = linear1(x) out = linear2(out) - fc1_w_mon1 = np.zeros((linear1.weight.shape)).astype( + fc1_w_mon1 = np.zeros(linear1.weight.shape).astype( "float32" ) - fc1_w_mon2 = np.zeros((linear1.weight.shape)).astype( + fc1_w_mon2 = np.zeros(linear1.weight.shape).astype( "float32" ) - fc1_b_mon1 = np.zeros((linear1.bias.shape)).astype( + fc1_b_mon1 = np.zeros(linear1.bias.shape).astype("float32") + fc1_b_mon2 = np.zeros(linear1.bias.shape).astype("float32") + fc2_w_mon1 = np.zeros(linear2.weight.shape).astype( "float32" ) - fc1_b_mon2 = np.zeros((linear1.bias.shape)).astype( - "float32" - ) - fc2_w_mon1 = np.zeros((linear2.weight.shape)).astype( - "float32" - ) - fc2_w_mon2 = np.zeros((linear2.weight.shape)).astype( - "float32" - ) - fc2_b_mon1 = np.zeros((linear2.bias.shape)).astype( - "float32" - ) - fc2_b_mon2 = np.zeros((linear2.bias.shape)).astype( + fc2_w_mon2 = np.zeros(linear2.weight.shape).astype( "float32" ) + fc2_b_mon1 = np.zeros(linear2.bias.shape).astype("float32") + fc2_b_mon2 = np.zeros(linear2.bias.shape).astype("float32") cost = paddle.nn.functional.square_error_cost( input=out, label=y diff --git a/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py index 388ace1ed33071ccccf4526a022c56644a0271e4..e171625dd4367d24d917018e36c6491da0eb3215 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py @@ -40,7 +40,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): self.op_type = "check_finite_and_unscale" self.init_dtype() x = np.random.random((8, 8)).astype(self.dtype) - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { 'FoundInfinite': np.array([0]), @@ -64,7 +64,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.nan x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -91,7 +91,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.nan x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) myscale = np.array([0.05]).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -118,7 +118,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.inf x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) myscale = np.array([0.05]).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py index 5067baf57c9993b963d0e2fdebc84a6ac296de02..ebd48f55d57f1273ad76974905af622a93a25fd6 100755 --- a/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py @@ -409,7 +409,7 @@ class XPUTestBilinearInterpV2Op(XPUOpTestWrapper): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py index 7b11ec3667527db8f1694b4336847033f90d0d16..3f1b1066b910fbc22bf13ae58fed812f5763a56f 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py @@ -96,7 +96,7 @@ class XPUTestCoalesceTensorOp(XPUOpTestWrapper): coalesce_tensor_var = np.concatenate([input for input in inputs]) if set_constant: coalesce_tensor_var = ( - np.ones((len(coalesce_tensor_var))) * constant + np.ones(len(coalesce_tensor_var)) * constant ) outputs = [ ( diff --git a/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py index 0bc1094888b0ae381fac857f33973a223523471d..872418812752224eccc9ecf209abe35748abba65 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py @@ -126,7 +126,7 @@ class XPUTestEmptyOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py index 4f7e9dd5e9f8ef7c8d228a8841b732160c8ee16c..f7098282a62a3847745479b2833d58d27da761c6 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py @@ -118,7 +118,7 @@ class XPUTestExpandV2Op(XPUOpTestWrapper): expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py index 086376f960cc179136d18a0413c751329606875b..4bd9abae9a5b908e2b7eb77bd1dcb75f299227de 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py @@ -147,7 +147,7 @@ class XPUTestFillConstantOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} @@ -178,7 +178,7 @@ class XPUTestFillConstantOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py index f1aac97dcba7a9a9a694e82b26d3f9f2b20fb64c..feb7549a33e500b806e7f46a446ba69b488dbc5f 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py @@ -265,16 +265,16 @@ class APITestStaticFusedFFN(unittest.TestCase): linear1_weight_data = np.random.random( (d_model, dim_feedforward) ).astype(dtype) - linear1_bias_data = np.zeros((dim_feedforward)).astype(dtype) + linear1_bias_data = np.zeros(dim_feedforward).astype(dtype) linear2_weight_data = np.random.random( (dim_feedforward, d_model) ).astype(dtype) - linear2_bias_data = np.zeros((d_model)).astype(dtype) + linear2_bias_data = np.zeros(d_model).astype(dtype) - ln1_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln1_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) - ln2_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln2_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) + ln1_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln1_bias_data = np.zeros(d_model).astype(layer_norm_dtype) + ln2_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln2_bias_data = np.zeros(d_model).astype(layer_norm_dtype) res_list = [fused_out, ln_out] real_res = [] diff --git a/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py index 3df6d2978c7bf97294e939996637c01b062c84a7..9d5b5e747f4458d666cded72607a5de9ccbf66f5 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py @@ -94,7 +94,7 @@ class XPUTestGaussianRandomOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.attrs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py index 7a9150312c9f7857c18d615765de44f70c1072f2..35c362b9a9aba17a842d82bf535979479836685d 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py @@ -470,7 +470,7 @@ class XPUNearestInterpOpWrapper(XPUOpTestWrapper): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py index 6e74b437e2d3131bbdf8630b095069406b871ebf..baeff8a10a6401a5700e829934665297f768bd74 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py @@ -35,7 +35,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-10, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob diff --git a/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py index 4ceaa44c0fbeb0fa05c989643f7f80a1cb774e4e..0e285f6b03c0e41ec1ebf5c4abe2ea55979ca31e 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py @@ -78,7 +78,7 @@ class XPUTestRandpermOp(XPUOpTestWrapper): self.use_xpu = True self.use_mkldnn = False self.inputs = {} - self.outputs = {"Out": np.zeros((self.n)).astype(self.dtype)} + self.outputs = {"Out": np.zeros(self.n).astype(self.dtype)} self.attrs = { "n": self.n, "dtype": convert_dtype(self.dtype), diff --git a/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py index e85ccf0cc44569747dc8a2b62ac44e2297f1ae66..41415a574b589540d7edf04c6e0c02a85d2b492a 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py @@ -118,7 +118,7 @@ class XPUTestReshapeOp(XPUOpTestWrapper): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py index c8d2d6ba2c846931a2dc210e5c81f0a7969fbd00..b5284dfb3208286c3852e17925ec73d98a4a1f87 100755 --- a/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py @@ -97,7 +97,7 @@ class XPUTestRNNOp(XPUOpTestWrapper): self.hidden_size, ) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -126,7 +126,7 @@ class XPUTestRNNOp(XPUOpTestWrapper): ('last_hidden', last_hidden), ('last_cell', last_cell), ], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py index bd0d3e9e3621142b894d9932ce252fdbc93f45ba..65f52bcfc0b1daf7ed3060db25fe36bec8000812 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py @@ -119,7 +119,7 @@ class TestSequenceUnpadOpError(unittest.TestCase): def test_length_variable(): x1 = paddle.static.data(name='x1', shape=[10, 5], dtype='float32') - len1 = np.random.random((10)).astype("int64") + len1 = np.random.random(10).astype("int64") paddle.static.nn.sequence_lod.sequence_unpad(x=x1, length=len1) self.assertRaises(TypeError, test_length_variable) diff --git a/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py index 7f0982c29cec31540e204afd5c10751fb1463d9c..c6f9c79be4d6f14cf8f4eb23ddcc8e1e886c1b87 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py @@ -110,7 +110,7 @@ class XPUTestTileOpRank1_tensor_attr(XPUOpTestWrapper): repeat_times_tensor = [] for index, ele in enumerate(self.repeat_times): repeat_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py index 53ab29a322a1138d862a2451f95675be719c34b4..56862299074f30b28f3835479ebe0cb4b5e01e53 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py @@ -113,7 +113,7 @@ class XPUTestUnsqueeze2Op(XPUOpTestWrapper): axes_tensor_list = [] for index, ele in enumerate(self.axes): axes_tensor_list.append( - ("axes" + str(index), np.ones((1)).astype('int32') * ele) + ("axes" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py index 82c974cd4fed954af19593013cbf1095235be8a7..a7a26f32b02cbee72b586476252765f3942e6ba1 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py @@ -47,7 +47,7 @@ class XPUTestWhereOp(XPUOpTestWrapper): def init_data(self): self.x = np.random.uniform(-3, 5, (100)).astype(self.dtype) self.y = np.random.uniform(-3, 5, (100)).astype(self.dtype) - self.cond = np.zeros((100)).astype("bool") + self.cond = np.zeros(100).astype("bool") def init_config(self): self.op_type = "where" diff --git a/python/paddle/hapi/model.py b/python/paddle/hapi/model.py index e25f88584ac37366769b0914c31f8e941e72e0c9..946766a82a1fc3ff05dc2ff8efbb1673e32bfff6 100644 --- a/python/paddle/hapi/model.py +++ b/python/paddle/hapi/model.py @@ -1514,7 +1514,7 @@ class Model: except ValueError as err: if skip_mismatch: warnings.warn( - ("Skip loading for {}. ".format(key) + str(err)) + "Skip loading for {}. ".format(key) + str(err) ) # reset optimizer when mismatch happens reset_optimizer = True diff --git a/python/paddle/nn/decode.py b/python/paddle/nn/decode.py index b5e1e287ecf6faa53d28cbdecdaed838dfb2bd42..04d40124016868a729bf386e4191d1d229edb223 100644 --- a/python/paddle/nn/decode.py +++ b/python/paddle/nn/decode.py @@ -706,7 +706,7 @@ def _dynamic_decode_imperative( initial_states, initial_finished, ) - cond = paddle.logical_not((paddle.all(initial_finished))) + cond = paddle.logical_not(paddle.all(initial_finished)) sequence_lengths = paddle.cast(paddle.zeros_like(initial_finished), "int64") outputs = None @@ -812,7 +812,7 @@ def _dynamic_decode_declarative( global_finished.stop_gradient = True step_idx = paddle.full(shape=[1], fill_value=0, dtype="int64") - cond = paddle.logical_not((paddle.all(initial_finished))) + cond = paddle.logical_not(paddle.all(initial_finished)) if max_step_num is not None: max_step_num = paddle.full( shape=[1], fill_value=max_step_num, dtype="int64" diff --git a/python/paddle/nn/layer/layers.py b/python/paddle/nn/layer/layers.py index b1f184469faadfabb2277a96d8c789e5dd426aad..23c02148c5bec726a35aa6fb0776b8f7b5af192a 100644 --- a/python/paddle/nn/layer/layers.py +++ b/python/paddle/nn/layer/layers.py @@ -1896,7 +1896,7 @@ class Layer: match_res = _check_match(key_name, param) matched_param_state.append(match_res) except ValueError as err: - warnings.warn(("Skip loading for {}. ".format(key) + str(err))) + warnings.warn("Skip loading for {}. ".format(key) + str(err)) for key in state_dict.keys(): if key not in match_keys: unexpected_keys.append(key) diff --git a/test/legacy_test/test_transforms_static.py b/test/legacy_test/test_transforms_static.py index 8e6c14635c20681381f590a615354b252c84eb7c..98bc132486c826979027021576c3f5d08e5ec500 100644 --- a/test/legacy_test/test_transforms_static.py +++ b/test/legacy_test/test_transforms_static.py @@ -126,7 +126,7 @@ class TestRandomCrop_random(TestTransformUnitTestBase): for y in range(h - c_h): for x in range(w - c_w): diff_abs_sum = np.abs( - (self.img[:, y : y + c_h, x : x + c_w] - res) + self.img[:, y : y + c_h, x : x + c_w] - res ).sum() if diff_abs_sum < eps: res_assert = False diff --git a/tools/get_pr_ut.py b/tools/get_pr_ut.py index 3377bf9f7fc0dcefdac3c2e8fd8a6406ef2e5023..91b3cb4d927dcde6cc062264c01beef696871c6d 100644 --- a/tools/get_pr_ut.py +++ b/tools/get_pr_ut.py @@ -162,7 +162,7 @@ class PRChecker: ) if 'cmakelist' in filename.lower(): isWhiteFile = False - elif filename.startswith((not_white_files)): + elif filename.startswith(not_white_files): isWhiteFile = False else: isWhiteFile = True @@ -330,7 +330,7 @@ class PRChecker: if filename.startswith(PADDLE_ROOT + 'python/'): file_list.append(filename) elif filename.startswith(PADDLE_ROOT + 'paddle/'): - if filename.startswith((PADDLE_ROOT + 'paddle/infrt')): + if filename.startswith(PADDLE_ROOT + 'paddle/infrt'): filterFiles.append(filename) elif filename.startswith(PADDLE_ROOT + 'paddle/scripts'): if filename.startswith( diff --git a/tools/get_single_test_cov.py b/tools/get_single_test_cov.py index 415d43e0b0109900c8b13073bb1ecbf8c5c2a38f..ba95a59fc1a81dbfd771d82c9e512a92ec7a74a6 100644 --- a/tools/get_single_test_cov.py +++ b/tools/get_single_test_cov.py @@ -48,8 +48,8 @@ def getFNDAFile(rootPath, test): os.system('echo %s >> %s' % (message_list[1], fn_filename)) if 'FNH:0' not in gcov_data: for message in message_list: - if message.startswith(('FNDA:')) and ( - not message.startswith(('FNDA:0,')) + if message.startswith('FNDA:') and ( + not message.startswith('FNDA:0,') ): tmp_data = message.split('FNDA:')[1].split(',') hit = int(tmp_data[0]) @@ -159,8 +159,8 @@ def getBaseFnda(rootPath, test): # only for cc file if ".cc" in message_list[1]: for message in message_list: - if message.startswith(('FNDA:')) and ( - not message.startswith(('FNDA:0,')) + if message.startswith('FNDA:') and ( + not message.startswith('FNDA:0,') ): tmp_data = message.split('FNDA:')[1].split(',') symbol_fnda[tmp_data[1]] = int(tmp_data[0])