From c069729680f60d42bc788b2e8e454b18c86cc69f Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?=E5=BC=A0=E6=98=A5=E4=B9=94?= <83450930+Liyulingyue@users.noreply.github.com> Date: Wed, 29 Mar 2023 15:30:56 +0800 Subject: [PATCH] [CodeStyle][UP034] remove (()) cases (#52060) * add up34 * modify var name in loop * revert changes in test_slice * Revert "modify var name in loop" This reverts commit 6d748e371afb417054ed0c6b36fd11e87959a90d. * temporarily ignore test_slice.py * add comment * empty commit, re-trigger all ci * fix inc --------- Co-authored-by: SigureMo --- .../operators/generator/cross_validate.py | 2 +- pyproject.toml | 7 +- .../auto_parallel/tuner/profiler.py | 2 +- .../meta_optimizers/ascend/ascend_parser.py | 4 +- python/paddle/distribution/normal.py | 2 +- .../test_distribution_transform.py | 4 +- .../test_basic_api_transformation.py | 2 +- .../dygraph_to_static/test_break_continue.py | 2 +- .../dygraph_to_static/test_cache_program.py | 6 +- .../unittests/dygraph_to_static/test_dict.py | 2 +- .../unittests/dygraph_to_static/test_list.py | 4 +- .../dygraph_to_static/test_return.py | 2 +- .../dygraph_to_static/test_transformer.py | 5 +- .../unittests/ipu/test_modelruntime_ipu.py | 6 +- ...est_conv_elementwise_add2_act_fuse_pass.py | 92 ++++++++----------- .../ir/inference/test_mkldnn_prelu_op.py | 4 +- .../ir/inference/test_trt_convert_where.py | 6 +- .../mkldnn/test_expand_v2_mkldnn_op.py | 2 +- .../mkldnn/test_fc_int8_mkldnn_op.py | 2 +- .../mkldnn/test_fill_constant_mkldnn_op.py | 2 +- .../unittests/mkldnn/test_matmul_mkldnn_op.py | 8 +- .../unittests/mkldnn/test_prelu_mkldnn_op.py | 4 +- .../mkldnn/test_reshape_mkldnn_op.py | 4 +- .../unittests/mkldnn/test_slice_mkldnn_op.py | 2 +- .../mkldnn/test_split_bf16_mkldnn_op.py | 2 +- .../unittests/mkldnn/test_split_mkldnn_op.py | 2 +- .../sequence/test_sequence_pad_op.py | 6 +- .../sequence/test_sequence_unpad_op.py | 2 +- .../tests/unittests/test_activation_op.py | 2 +- .../fluid/tests/unittests/test_adamw_op.py | 16 ++-- .../fluid/tests/unittests/test_addmm_op.py | 10 +- .../test_amp_check_finite_and_scale_op.py | 6 +- .../tests/unittests/test_bernoulli_op.py | 2 +- .../unittests/test_bilinear_interp_op.py | 2 +- .../unittests/test_bilinear_interp_v2_op.py | 2 +- .../tests/unittests/test_box_coder_op.py | 4 +- .../unittests/test_coalesce_tensor_op.py | 2 +- .../fluid/tests/unittests/test_compare_op.py | 2 +- .../tests/unittests/test_complex_matmul.py | 4 +- .../tests/unittests/test_crop_tensor_op.py | 4 +- .../unittests/test_cross_entropy_loss.py | 2 +- .../fluid/tests/unittests/test_cumsum_op.py | 2 +- .../unittests/test_cumulative_trapezoid.py | 2 +- .../fluid/tests/unittests/test_dropout_op.py | 2 +- .../fluid/tests/unittests/test_eig_op.py | 2 +- .../unittests/test_elementwise_gradient_op.py | 2 +- .../fluid/tests/unittests/test_empty_op.py | 2 +- .../fluid/tests/unittests/test_expand_op.py | 2 +- .../tests/unittests/test_expand_v2_op.py | 2 +- .../fluid/tests/unittests/test_fc_op.py | 2 +- .../tests/unittests/test_fill_constant_op.py | 4 +- .../tests/unittests/test_full_like_op.py | 4 +- .../unittests/test_fused_feedforward_op.py | 12 +-- .../unittests/test_fused_gate_attention_op.py | 2 +- .../unittests/test_gaussian_random_op.py | 2 +- .../fluid/tests/unittests/test_gru_rnn_op.py | 4 +- .../tests/unittests/test_gumbel_softmax_op.py | 2 +- .../unittests/test_imperative_framework.py | 2 +- .../fluid/tests/unittests/test_initializer.py | 4 +- .../tests/unittests/test_initializer_nn.py | 2 +- .../tests/unittests/test_jit_save_load.py | 8 +- .../fluid/tests/unittests/test_linear.py | 2 +- .../tests/unittests/test_linear_interp_op.py | 2 +- .../unittests/test_linear_interp_v2_op.py | 2 +- .../tests/unittests/test_lstm_cudnn_op.py | 4 +- .../fluid/tests/unittests/test_meshgrid_op.py | 2 +- .../tests/unittests/test_multi_dot_op.py | 18 ++-- .../fluid/tests/unittests/test_mv_op.py | 4 +- .../fluid/tests/unittests/test_nanmedian.py | 2 +- .../tests/unittests/test_nearest_interp_op.py | 2 +- .../unittests/test_nearest_interp_v2_op.py | 2 +- .../tests/unittests/test_npair_loss_op.py | 2 +- .../fluid/tests/unittests/test_numel_op.py | 4 +- .../fluid/tests/unittests/test_pool2d_op.py | 2 +- .../fluid/tests/unittests/test_pylayer_op.py | 8 +- .../fluid/tests/unittests/test_randint_op.py | 6 +- .../fluid/tests/unittests/test_randperm_op.py | 2 +- .../tests/unittests/test_rank_attention_op.py | 2 +- .../fluid/tests/unittests/test_reshape_op.py | 2 +- .../tests/unittests/test_rnn_cell_api.py | 2 +- .../fluid/tests/unittests/test_rnn_op.py | 4 +- .../fluid/tests/unittests/test_seed_op.py | 4 +- .../unittests/test_sigmoid_focal_loss_op.py | 4 +- .../tests/unittests/test_simple_rnn_op.py | 4 +- .../fluid/tests/unittests/test_slice_op.py | 6 +- .../fluid/tests/unittests/test_split_op.py | 2 +- .../tests/unittests/test_strided_slice_op.py | 6 +- .../fluid/tests/unittests/test_tile_op.py | 2 +- .../fluid/tests/unittests/test_trapezoid.py | 2 +- .../unittests/test_trilinear_interp_op.py | 2 +- .../unittests/test_trilinear_interp_v2_op.py | 2 +- .../unittests/test_uniform_random_bf16_op.py | 2 +- .../tests/unittests/test_uniform_random_op.py | 12 +-- .../tests/unittests/test_unsqueeze2_op.py | 2 +- .../fluid/tests/unittests/test_variable.py | 2 +- .../fluid/tests/unittests/test_yolo_box_op.py | 14 +-- .../tests/unittests/test_yolov3_loss_op.py | 2 +- .../unittests/xpu/test_activation_op_xpu.py | 2 +- .../tests/unittests/xpu/test_adamw_op_xpu.py | 24 ++--- .../test_amp_check_finite_and_scale_op_xpu.py | 8 +- .../xpu/test_bilinear_interp_v2_op_xpu.py | 2 +- .../xpu/test_coalesce_tensor_op_xpu.py | 2 +- .../tests/unittests/xpu/test_empty_op_xpu.py | 2 +- .../unittests/xpu/test_expand_v2_op_xpu.py | 2 +- .../xpu/test_fill_constant_op_xpu.py | 4 +- .../xpu/test_fused_feedforward_op_xpu.py | 12 +-- .../xpu/test_gaussian_random_op_xpu.py | 2 +- .../xpu/test_nearest_interp_v2_op_xpu.py | 2 +- .../unittests/xpu/test_randint_op_xpu.py | 2 +- .../unittests/xpu/test_randperm_op_xpu.py | 2 +- .../unittests/xpu/test_reshape2_op_xpu.py | 2 +- .../tests/unittests/xpu/test_rnn_op_xpu.py | 4 +- .../xpu/test_sequence_unpad_op_xpu.py | 2 +- .../tests/unittests/xpu/test_tile_op_xpu.py | 2 +- .../unittests/xpu/test_unsqueeze2_op_xpu.py | 2 +- .../tests/unittests/xpu/test_where_op_xpu.py | 2 +- python/paddle/hapi/model.py | 2 +- python/paddle/nn/decode.py | 4 +- python/paddle/nn/layer/layers.py | 2 +- test/legacy_test/test_transforms_static.py | 2 +- tools/get_pr_ut.py | 4 +- tools/get_single_test_cov.py | 8 +- 122 files changed, 266 insertions(+), 288 deletions(-) diff --git a/paddle/fluid/operators/generator/cross_validate.py b/paddle/fluid/operators/generator/cross_validate.py index 82d21f47c62..cfe013e73e4 100644 --- a/paddle/fluid/operators/generator/cross_validate.py +++ b/paddle/fluid/operators/generator/cross_validate.py @@ -26,7 +26,7 @@ def main(forward_op_yaml_paths, backward_op_yaml_paths): with open(op_yaml_path, "rt", encoding="utf-8") as f: op_list = yaml.safe_load(f) if op_list is not None: - ops.update(to_named_dict((op_list))) + ops.update(to_named_dict(op_list)) cross_validate(ops) diff --git a/pyproject.toml b/pyproject.toml index 75bd21e7ba9..6519ad98c6f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -80,7 +80,7 @@ select = [ # "UP031", # "UP032", "UP033", - # "UP034", + "UP034", "UP035", "UP036", @@ -168,6 +168,11 @@ unfixable = [ ] [tool.ruff.per-file-ignores] +# Ignore unused imports in __init__.py "__init__.py" = ["F401"] +# Temporarily ignore test_slice.py to avoid PR-CI-CINN failure, please fix! +"python/paddle/fluid/tests/unittests/dygraph_to_static/test_slice.py" = ["UP034"] +# Ignore version check in setup.py "setup.py" = ["UP036"] +# Ignore unnecessary lambda in dy2st unittest test_lambda "python/paddle/fluid/tests/unittests/dygraph_to_static/test_lambda.py" = ["PLC3002"] diff --git a/python/paddle/distributed/auto_parallel/tuner/profiler.py b/python/paddle/distributed/auto_parallel/tuner/profiler.py index c419ae7fcf7..138a8bddd41 100644 --- a/python/paddle/distributed/auto_parallel/tuner/profiler.py +++ b/python/paddle/distributed/auto_parallel/tuner/profiler.py @@ -270,7 +270,7 @@ def profiler(args): with open(result_path, 'w') as fp: json.dump(result_dict, fp) - print("profile done! avg speed : {} step / s.".format((avg_tput))) + print("profile done! avg speed : {} step / s.".format(avg_tput)) except paddle.framework.core.EOFException: data_loader._inner_dataloader.reset() diff --git a/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py b/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py index 99a30333ac7..21f3e26da74 100644 --- a/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py +++ b/python/paddle/distributed/fleet/meta_optimizers/ascend/ascend_parser.py @@ -219,7 +219,7 @@ class AscendParserBase: tensor = core.GETensor(tensor_desc) data = ( - (value * np.ones((shape))) + (value * np.ones(shape)) .reshape(shape) .astype(self.ascend_helper.dtype2np(dtype)) ) @@ -282,7 +282,7 @@ class AscendParserBase: ) tensor = core.GETensor(tensor_desc) - data = np.ones((2)).astype("int32").reshape([2]) + data = np.ones(2).astype("int32").reshape([2]) data[0] = 64 buf = data.tobytes() data_8 = np.frombuffer(buf, dtype=np.uint8) diff --git a/python/paddle/distribution/normal.py b/python/paddle/distribution/normal.py index 22ebab0ed4b..edb08f6cfae 100644 --- a/python/paddle/distribution/normal.py +++ b/python/paddle/distribution/normal.py @@ -241,7 +241,7 @@ class Normal(distribution.Distribution): ) return paddle.add( 0.5 + zero_tmp, - 0.5 * math.log(2 * math.pi) + paddle.log((self.scale + zero_tmp)), + 0.5 * math.log(2 * math.pi) + paddle.log(self.scale + zero_tmp), name=name, ) diff --git a/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py b/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py index 729ae93d647..4bc90d8f792 100644 --- a/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py +++ b/python/paddle/fluid/tests/unittests/distribution/test_distribution_transform.py @@ -1185,7 +1185,7 @@ class TestStickBreakingTransform(unittest.TestCase): def test_codomain(self): self.assertTrue(isinstance(self._t._codomain, variable.Variable)) - @param.param_func(((np.random.random((10)),),)) + @param.param_func(((np.random.random(10),),)) def test_forward(self, input): np.testing.assert_allclose( self._t.inverse(self._t.forward(paddle.to_tensor(input))), @@ -1202,7 +1202,7 @@ class TestStickBreakingTransform(unittest.TestCase): def test_inverse_shape(self, shape, expected_shape): self.assertEqual(self._t.inverse_shape(shape), expected_shape) - @param.param_func(((np.random.random((10)),),)) + @param.param_func(((np.random.random(10),),)) def test_forward_log_det_jacobian(self, x): self.assertEqual( self._t.forward_log_det_jacobian(paddle.to_tensor(x)).shape, [1] diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py index c3bbd2b7865..ea6cb541ff9 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_basic_api_transformation.py @@ -456,7 +456,7 @@ def _dygraph_fn(): x = np.random.random((1, 3)).astype('float32') with fluid.dygraph.guard(): fluid.dygraph.to_variable(x) - np.random.random((1)) + np.random.random(1) class TestDygraphApiRecognition(unittest.TestCase): diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py index e1d9185f57f..7a4340d805a 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_break_continue.py @@ -204,7 +204,7 @@ def test_optim_break_in_while(x): class TestContinueInFor(unittest.TestCase): def setUp(self): - self.input = np.zeros((1)).astype('int64') + self.input = np.zeros(1).astype('int64') self.place = ( fluid.CUDAPlace(0) if fluid.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py index ae7a0812799..69100e60efb 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_cache_program.py @@ -138,7 +138,7 @@ class TestConvertWithCache(unittest.TestCase): @to_static def sum_even_until_limit(max_len, limit): - ret_sum = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) + ret_sum = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) for i in range(max_len): if i % 2 > 0: continue @@ -150,8 +150,8 @@ def sum_even_until_limit(max_len, limit): def sum_under_while(limit): - i = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) - ret_sum = fluid.dygraph.to_variable(np.zeros((1)).astype('int32')) + i = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) + ret_sum = fluid.dygraph.to_variable(np.zeros(1).astype('int32')) while i <= limit: ret_sum += i i += 1 diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py index 0aeff4961fb..852b69e1a7b 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_dict.py @@ -171,7 +171,7 @@ def test_dic_pop_2(x): class TestDictPop(unittest.TestCase): def setUp(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') self.place = ( paddle.CUDAPlace(0) if paddle.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py index cecfb00266a..44e02950bc5 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_list.py @@ -219,7 +219,7 @@ class TestListWithoutControlFlow(unittest.TestCase): self.init_dygraph_func() def init_data(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') def init_dygraph_func(self): self.all_dygraph_funcs = [ @@ -275,7 +275,7 @@ class TestListInIf(TestListWithoutControlFlow): class TestListInWhileLoop(TestListWithoutControlFlow): def init_data(self): - self.input = np.random.random((3)).astype('int32') + self.input = np.random.random(3).astype('int32') self.iter_num = 3 def init_dygraph_func(self): diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py index 0f72885fca9..2796cc332df 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_return.py @@ -265,7 +265,7 @@ def test_return_nested(x): class TestReturnBase(unittest.TestCase): def setUp(self): - self.input = np.ones((1)).astype('int32') + self.input = np.ones(1).astype('int32') self.place = ( fluid.CUDAPlace(0) if fluid.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py index 459b3719d2d..47dadf6f71a 100644 --- a/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py +++ b/python/paddle/fluid/tests/unittests/dygraph_to_static/test_transformer.py @@ -108,7 +108,7 @@ def train_static(args, batch_generator): # the best cross-entropy value with label smoothing loss_normalizer = -( (1.0 - args.label_smooth_eps) - * np.log((1.0 - args.label_smooth_eps)) + * np.log(1.0 - args.label_smooth_eps) + args.label_smooth_eps * np.log( args.label_smooth_eps / (args.trg_vocab_size - 1) + 1e-20 @@ -221,8 +221,7 @@ def train_dygraph(args, batch_generator): ) # the best cross-entropy value with label smoothing loss_normalizer = -( - (1.0 - args.label_smooth_eps) - * np.log((1.0 - args.label_smooth_eps)) + (1.0 - args.label_smooth_eps) * np.log(1.0 - args.label_smooth_eps) + args.label_smooth_eps * np.log(args.label_smooth_eps / (args.trg_vocab_size - 1) + 1e-20) ) diff --git a/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py b/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py index 46b03297cd1..5e8489c1f5d 100644 --- a/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py +++ b/python/paddle/fluid/tests/unittests/ipu/test_modelruntime_ipu.py @@ -123,10 +123,8 @@ class TestBase(IPUOpTest): pad_batch = self.batch_size - dy_batch for k, v in feed.items(): pad_size = tuple( - ( - (0, 0 if i != 0 else pad_batch) - for i in range(len(v.shape)) - ) + (0, 0 if i != 0 else pad_batch) + for i in range(len(v.shape)) ) feed[k] = np.pad(v, pad_size, 'constant', constant_values=0) diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py b/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py index 6b30f83ae0c..397933737ba 100755 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_conv_elementwise_add2_act_fuse_pass.py @@ -62,24 +62,20 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): if ( int( ( - ( - input_shape[2] - - (dilations[0] * (filter_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + input_shape[2] + - (dilations[0] * (filter_shape[2] - 1) + 1) ) + / strides[0] + + 1 ) <= 0 or int( ( - ( - input_shape[3] - - (dilations[1] * (filter_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + input_shape[3] + - (dilations[1] * (filter_shape[3] - 1) + 1) ) + / strides[1] + + 1 ) <= 0 ): @@ -88,28 +84,24 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): if ( int( ( - ( - input_shape[2] - + paddings[0] - + paddings[1] - - (dilations[0] * (filter_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + input_shape[2] + + paddings[0] + + paddings[1] + - (dilations[0] * (filter_shape[2] - 1) + 1) ) + / strides[0] + + 1 ) <= 0 or int( ( - ( - input_shape[3] - + paddings[2] - + paddings[3] - - (dilations[1] * (filter_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + input_shape[3] + + paddings[2] + + paddings[3] + - (dilations[1] * (filter_shape[3] - 1) + 1) ) + / strides[1] + + 1 ) <= 0 ): @@ -206,27 +198,23 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): f_shape[0], int( ( - ( - x_shape[2] - + padding[0] - + padding[1] - - (dilations[0] * (f_shape[2] - 1) + 1) - ) - / strides[0] - + 1 + x_shape[2] + + padding[0] + + padding[1] + - (dilations[0] * (f_shape[2] - 1) + 1) ) + / strides[0] + + 1 ), int( ( - ( - x_shape[3] - + padding[2] - + padding[3] - - (dilations[1] * (f_shape[3] - 1) + 1) - ) - / strides[1] - + 1 + x_shape[3] + + padding[2] + + padding[3] + - (dilations[1] * (f_shape[3] - 1) + 1) ) + / strides[1] + + 1 ), ] @@ -241,18 +229,14 @@ class TestConvElementwiseAdd2ActPass(PassAutoScanTest): x_shape[0], f_shape[0], int( - ( - (x_shape[2] - (dilations[0] * (f_shape[2] - 1) + 1)) - / strides[0] - + 1 - ) + (x_shape[2] - (dilations[0] * (f_shape[2] - 1) + 1)) + / strides[0] + + 1 ), int( - ( - (x_shape[3] - (dilations[1] * (f_shape[3] - 1) + 1)) - / strides[1] - + 1 - ) + (x_shape[3] - (dilations[1] * (f_shape[3] - 1) + 1)) + / strides[1] + + 1 ), ] bias_index = 1 diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py b/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py index 1ad4600df3e..cab24fb2217 100644 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_mkldnn_prelu_op.py @@ -42,7 +42,7 @@ class TestMkldnnPreluOp(MkldnnAutoScanTest): elif kwargs["mode"] == "channel": if len(kwargs['in_shape']) <= 1: # not valid case, just return 0 - return np.zeros((1)).astype(np.float32) + return np.zeros(1).astype(np.float32) if kwargs['data_format'] == 'NCHW': return np.random.random(kwargs['in_shape'][1]).astype( np.float32 @@ -54,7 +54,7 @@ class TestMkldnnPreluOp(MkldnnAutoScanTest): else: if len(kwargs['in_shape']) <= 1: # not valid case, just return 0 - return np.zeros((1)).astype(np.float32) + return np.zeros(1).astype(np.float32) return np.random.random(kwargs['in_shape']).astype(np.float32) prelu_op = OpConfig( diff --git a/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py b/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py index 269a8edae19..17c20e56de2 100644 --- a/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py +++ b/python/paddle/fluid/tests/unittests/ir/inference/test_trt_convert_where.py @@ -35,7 +35,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input1(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: @@ -45,7 +45,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input2(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: @@ -55,7 +55,7 @@ class TrtConvertActivationTest(TrtLayerAutoScanTest): def generate_input3(dims, batch): if dims == 1: - return np.zeros((batch)).astype(np.float32) + return np.zeros(batch).astype(np.float32) elif dims == 2: return np.ones((batch, 4)).astype(np.float32) elif dims == 3: diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py index 1747184eaf9..2d500a2e4fb 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_expand_v2_mkldnn_op.py @@ -90,7 +90,7 @@ class TestExpandV2ExpandShapesTensor1OneDNNOp(TestExpandV2OneDNNOp): self.expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): self.expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) def set_additional_inputs(self): diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py index ae68c3f5bbe..a6852e30819 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_fc_int8_mkldnn_op.py @@ -64,7 +64,7 @@ class TestFCINT8OneDNNOp(OpTest): self.out_float = np.dot(self.x_float, self.y_float) if self.use_bias: - self.bias = np.random.random((10)).astype("float32") * 10 + self.bias = np.random.random(10).astype("float32") * 10 self.out_float += self.bias self.out_scale, self.out = self.quantize(self.out_float) diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py index fe1ca64dbe2..50cf635b28a 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_fill_constant_mkldnn_op.py @@ -78,7 +78,7 @@ class TestFillZerosLike4DShapeTensorListPriorityOneDNNOp( self.shape_tensor_list = [] for index, elem in enumerate(shape): self.shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * elem) + ("x" + str(index), np.ones(1).astype('int32') * elem) ) self.inputs = {'ShapeTensorList': self.shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py index 07e7334fcd5..07c4a5bf7de 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_matmul_mkldnn_op.py @@ -119,17 +119,17 @@ class TestDnnlMatMulOpMixedDimsXWiderTransposeX(TestDnnlMatMulWithGradOp): class TestDnnlMatMulOpVectorMultiply(TestDnnlMatMulWithGradOp): def generate_data(self): - self.x = np.random.random((5)).astype("float32") - self.y = np.random.random((5)).astype("float32") + self.x = np.random.random(5).astype("float32") + self.y = np.random.random(5).astype("float32") self.out = np.matmul(self.x, self.y) class TestDnnlMatMulOpVectorMultiplyTranspose(TestDnnlMatMulWithGradOp): def generate_data(self): - self.x = np.random.random((5)).astype("float32") + self.x = np.random.random(5).astype("float32") x_resized = np.copy(self.x) x_resized = np.expand_dims(x_resized, 1) - self.y = np.random.random((6)).astype("float32") + self.y = np.random.random(6).astype("float32") y_resized = np.copy(self.y) y_resized = np.expand_dims(y_resized, 0) self.out = np.matmul(x_resized, y_resized) diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py index c70fbd33204..13f1d0a7be7 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_prelu_mkldnn_op.py @@ -103,14 +103,14 @@ class TestPReluModeChannelAlpha1DOneDNNOp(TestPReluModeChannelOneDNNOp): def init_attrs(self): self.mode = "channel" self.x = np.random.random((1, 100, 1)).astype("float32") - self.alpha = np.random.random((100)).astype("float32") + self.alpha = np.random.random(100).astype("float32") class TestPReluModeAllAlpha1DOneDNNOp(TestPReluModeAllOneDNNOp): def init_attrs(self): self.mode = "channel" self.x = np.random.random((1, 1, 100)).astype("float32") - self.alpha = np.random.random((1)).astype("float32") + self.alpha = np.random.random(1).astype("float32") # BF16 TESTS diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py index 7d16dec8583..8e607d5a448 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_reshape_mkldnn_op.py @@ -119,7 +119,7 @@ class TestReshape2OneDNNOpDimInfer1_attr_ShapeTensor(TestReshape2OneDNNOp): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs["ShapeTensor"] = shape_tensor @@ -138,7 +138,7 @@ class TestReshape2OneDNNOpDimInfer1_attr_ShapeTensorAndShape( shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs["Shape"] = np.array((1, 2, 3, 4), dtype="int32") diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py index 59d6a22a680..b33ac0e34a0 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_slice_mkldnn_op.py @@ -160,7 +160,7 @@ class TestSliceOneDNNOp_decs_dim_starts_ListTensor( def set_inputs(self): starts_tensor = [] for index, ele in enumerate(self.starts): - starts_tensor.append(("x1", np.ones((1)).astype('int32') * 2)) + starts_tensor.append(("x1", np.ones(1).astype('int32') * 2)) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} def config(self): diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py index 38292215563..88fbdbde6a3 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_split_bf16_mkldnn_op.py @@ -96,7 +96,7 @@ class TestSplitSectionsTensorBF16OneDNNOp(TestSplitSectionsBF16OneDNNOp): self.sections_tensor_list = [] for index, ele in enumerate(self.sections): self.sections_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.sections = [-1, -1, -1] indices_or_sections = [2, 3] # sections diff --git a/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py b/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py index 88bedbb61cc..c1105b2096b 100644 --- a/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py +++ b/python/paddle/fluid/tests/unittests/mkldnn/test_split_mkldnn_op.py @@ -114,7 +114,7 @@ class TestSplitSectionsTensorOneDNNOp(TestSplitSectionsOneDNNOp): self.sections_tensor_list = [] for index, ele in enumerate(self.sections): self.sections_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.sections = [-1, -1, -1] indices_or_sections = [2, 3] # sections diff --git a/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py b/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py index 3dc34f8d099..64048753c3c 100644 --- a/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py +++ b/python/paddle/fluid/tests/unittests/sequence/test_sequence_pad_op.py @@ -70,7 +70,7 @@ class TestSequencePadOp(OpTest): start_idx = end_idx out_data = np.array(padded_sequences) - length = np.array(self.x_len_lod[0]).reshape((-1)) + length = np.array(self.x_len_lod[0]).reshape(-1) self.outputs = {'Out': out_data, 'Length': length} def setUp(self): @@ -90,7 +90,7 @@ class TestSequencePadOp2(TestSequencePadOp): def set_attr(self): self.x_shape = [12, 10] self.x_len_lod = [[2, 3, 4, 3]] - self.pad_value = np.random.random((10)) + self.pad_value = np.random.random(10) self.padded_length = -1 self.dtype = 'float64' @@ -108,7 +108,7 @@ class TestSequencePadOp4(TestSequencePadOp): def set_attr(self): self.x_shape = [12, 10] self.x_len_lod = [[2, 3, 4, 3]] - self.pad_value = np.random.random((10)) + self.pad_value = np.random.random(10) self.padded_length = 7 self.dtype = 'float64' diff --git a/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py b/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py index 930c25a2de4..daaa4b1ee9b 100644 --- a/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py +++ b/python/paddle/fluid/tests/unittests/sequence/test_sequence_unpad_op.py @@ -98,7 +98,7 @@ class TestSequenceUnpadOpError(unittest.TestCase): def test_length_variable(): x1 = paddle.static.data(name='x1', shape=[10, 5], dtype='float32') - len1 = np.random.random((10)).astype("int64") + len1 = np.random.random(10).astype("int64") paddle.static.nn.sequence_lod.sequence_pad(x=x1, length=len1) self.assertRaises(TypeError, test_length_variable) diff --git a/python/paddle/fluid/tests/unittests/test_activation_op.py b/python/paddle/fluid/tests/unittests/test_activation_op.py index 1badfb8e3f7..dfe0efabcc4 100644 --- a/python/paddle/fluid/tests/unittests/test_activation_op.py +++ b/python/paddle/fluid/tests/unittests/test_activation_op.py @@ -2444,7 +2444,7 @@ class TestSoftRelu(TestActivation): t = np.copy(x) t[t < -threshold] = -threshold t[t > threshold] = threshold - out = np.log((np.exp(t) + 1)) + out = np.log(np.exp(t) + 1) self.inputs = {'X': OpTest.np_dtype_to_fluid_dtype(x)} self.attrs = {'threshold': threshold} diff --git a/python/paddle/fluid/tests/unittests/test_adamw_op.py b/python/paddle/fluid/tests/unittests/test_adamw_op.py index ea72c703f23..e94a0de7e41 100644 --- a/python/paddle/fluid/tests/unittests/test_adamw_op.py +++ b/python/paddle/fluid/tests/unittests/test_adamw_op.py @@ -794,14 +794,14 @@ class TestAdamWOpLayerwiseLR(TestAdamWOp): out = linear1(x) out = linear2(out) - fc1_w_mon1 = np.zeros((linear1.weight.shape)).astype("float32") - fc1_w_mon2 = np.zeros((linear1.weight.shape)).astype("float32") - fc1_b_mon1 = np.zeros((linear1.bias.shape)).astype("float32") - fc1_b_mon2 = np.zeros((linear1.bias.shape)).astype("float32") - fc2_w_mon1 = np.zeros((linear2.weight.shape)).astype("float32") - fc2_w_mon2 = np.zeros((linear2.weight.shape)).astype("float32") - fc2_b_mon1 = np.zeros((linear2.bias.shape)).astype("float32") - fc2_b_mon2 = np.zeros((linear2.bias.shape)).astype("float32") + fc1_w_mon1 = np.zeros(linear1.weight.shape).astype("float32") + fc1_w_mon2 = np.zeros(linear1.weight.shape).astype("float32") + fc1_b_mon1 = np.zeros(linear1.bias.shape).astype("float32") + fc1_b_mon2 = np.zeros(linear1.bias.shape).astype("float32") + fc2_w_mon1 = np.zeros(linear2.weight.shape).astype("float32") + fc2_w_mon2 = np.zeros(linear2.weight.shape).astype("float32") + fc2_b_mon1 = np.zeros(linear2.bias.shape).astype("float32") + fc2_b_mon2 = np.zeros(linear2.bias.shape).astype("float32") cost = paddle.nn.functional.square_error_cost( input=out, label=y diff --git a/python/paddle/fluid/tests/unittests/test_addmm_op.py b/python/paddle/fluid/tests/unittests/test_addmm_op.py index 4e49a844d06..3041841cdf8 100644 --- a/python/paddle/fluid/tests/unittests/test_addmm_op.py +++ b/python/paddle/fluid/tests/unittests/test_addmm_op.py @@ -230,7 +230,7 @@ class TestAddMMOp4(OpTest): self.dtype = np.float64 self.init_dtype_type() self.inputs = { - 'Input': np.random.random((100)).astype(self.dtype), + 'Input': np.random.random(100).astype(self.dtype), 'X': np.random.random((20, 10)).astype(self.dtype), 'Y': np.random.random((10, 100)).astype(self.dtype), } @@ -296,7 +296,7 @@ class TestAddMMAPI(unittest.TestCase): self.assertRaises(ValueError, test_error1) def test_error2(): - data_x_wrong = np.ones((2)).astype(np.float32) + data_x_wrong = np.ones(2).astype(np.float32) x = paddle.to_tensor(data_x_wrong) y = paddle.to_tensor(data_y) input = paddle.to_tensor(data_input) @@ -318,7 +318,7 @@ class TestAddMMAPI(unittest.TestCase): self.assertRaises(ValueError, test_error3) def test_error4(): - data_input_wrong = np.ones((5)).astype(np.float32) + data_input_wrong = np.ones(5).astype(np.float32) x = paddle.to_tensor(data_x) y = paddle.to_tensor(data_y) input = paddle.to_tensor(data_input_wrong) @@ -358,7 +358,7 @@ class TestAddMMAPI(unittest.TestCase): def test_api_normal_2(self): data_x = np.ones((3, 10)).astype(np.float32) data_y = np.ones((10, 3)).astype(np.float32) - data_input = np.ones((3)).astype(np.float32) + data_input = np.ones(3).astype(np.float32) data_alpha = 0.1 data_beta = 1.0 @@ -383,7 +383,7 @@ class TestAddMMAPI(unittest.TestCase): def test_api_normal_3(self): data_x = np.ones((3, 10)).astype(np.float32) data_y = np.ones((10, 3)).astype(np.float32) - data_input = np.ones((1)).astype(np.float32) + data_input = np.ones(1).astype(np.float32) data_alpha = 0.1 data_beta = 1.0 diff --git a/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py b/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py index 11303d6e1a0..4926b10b4a1 100644 --- a/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py +++ b/python/paddle/fluid/tests/unittests/test_amp_check_finite_and_scale_op.py @@ -32,7 +32,7 @@ class TestCheckFiniteAndUnscaleOp(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] self.init_dtype() x = np.random.random((1024, 1024)).astype(self.dtype) - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -55,7 +55,7 @@ class TestCheckFiniteAndUnscaleOpWithNan(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] x = np.random.random((1024, 1024)).astype(self.dtype) x[128][128] = np.nan - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -80,7 +80,7 @@ class TestCheckFiniteAndUnscaleOpWithInf(OpTest): self.python_out_sig = ["out0", "FoundInfinite"] x = np.random.random((1024, 1024)).astype(self.dtype) x[128][128] = np.inf - scale = np.random.random((1)).astype(self.dtype) + scale = np.random.random(1).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { diff --git a/python/paddle/fluid/tests/unittests/test_bernoulli_op.py b/python/paddle/fluid/tests/unittests/test_bernoulli_op.py index 33a450310d1..af08b07237f 100644 --- a/python/paddle/fluid/tests/unittests/test_bernoulli_op.py +++ b/python/paddle/fluid/tests/unittests/test_bernoulli_op.py @@ -24,7 +24,7 @@ def output_hist(out): hist, _ = np.histogram(out, bins=2) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.5 * np.ones((2)) + prob = 0.5 * np.ones(2) return hist, prob diff --git a/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py b/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py index f4b051e1451..33104d3cf2d 100755 --- a/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_bilinear_interp_op.py @@ -446,7 +446,7 @@ class TestBilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py index 1b3f615fde3..b33e84bac87 100755 --- a/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_bilinear_interp_v2_op.py @@ -798,7 +798,7 @@ class TestBilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_box_coder_op.py b/python/paddle/fluid/tests/unittests/test_box_coder_op.py index 889026aeb30..c0beed3ca4a 100644 --- a/python/paddle/fluid/tests/unittests/test_box_coder_op.py +++ b/python/paddle/fluid/tests/unittests/test_box_coder_op.py @@ -293,7 +293,7 @@ class TestBoxCoderOpWithVariance(OpTest): self.python_api = wrapper_box_coder lod = [[1, 1, 1, 1, 1]] prior_box = np.random.random((30, 4)).astype('float32') - prior_box_var = np.random.random((4)).astype('float32') + prior_box_var = np.random.random(4).astype('float32') target_box = np.random.random((30, 81, 4)).astype('float32') code_type = "DecodeCenterSize" box_normalized = False @@ -325,7 +325,7 @@ class TestBoxCoderOpWithVarianceDygraphAPI(unittest.TestCase): def setUp(self): self.lod = [[1, 1, 1, 1, 1]] self.prior_box = np.random.random((30, 4)).astype('float32') - self.prior_box_var = np.random.random((4)).astype('float32') + self.prior_box_var = np.random.random(4).astype('float32') self.target_box = np.random.random((30, 81, 4)).astype('float32') self.code_type = "DecodeCenterSize" self.box_normalized = False diff --git a/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py b/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py index a75eb0f77a5..faf33532c33 100644 --- a/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py +++ b/python/paddle/fluid/tests/unittests/test_coalesce_tensor_op.py @@ -106,7 +106,7 @@ class TestAllocContinuousSpace(OpTest): coalesce_tensor_var = np.concatenate([input for input in inputs]) if set_constant: - coalesce_tensor_var = np.ones((len(coalesce_tensor_var))) * constant + coalesce_tensor_var = np.ones(len(coalesce_tensor_var)) * constant outputs = [ (out[0], np.ones(out[1].shape).astype(self.dtype) * constant) for out in outputs diff --git a/python/paddle/fluid/tests/unittests/test_compare_op.py b/python/paddle/fluid/tests/unittests/test_compare_op.py index 1ddc402c67b..d7dc124c5a3 100755 --- a/python/paddle/fluid/tests/unittests/test_compare_op.py +++ b/python/paddle/fluid/tests/unittests/test_compare_op.py @@ -332,7 +332,7 @@ def create_paddle_case(op_type, callback): op = eval("paddle.%s" % (self.op_type)) out = op(x, y) exe = paddle.static.Executor(self.place) - input_x = np.arange(0, 5).reshape((5)).astype(np.int32) + input_x = np.arange(0, 5).reshape(5).astype(np.int32) input_y = np.array([5, 3, 2]).reshape((3, 1)).astype(np.int32) real_result = callback(input_x, input_y) (res,) = exe.run( diff --git a/python/paddle/fluid/tests/unittests/test_complex_matmul.py b/python/paddle/fluid/tests/unittests/test_complex_matmul.py index 65924da12b0..c6a89bb3cec 100644 --- a/python/paddle/fluid/tests/unittests/test_complex_matmul.py +++ b/python/paddle/fluid/tests/unittests/test_complex_matmul.py @@ -110,8 +110,8 @@ class TestComplexMatMulLayer(unittest.TestCase): x = np.random.random((2, 1, 100)).astype( dtype ) + 1j * np.random.random((2, 1, 100)).astype(dtype) - y = np.random.random((100)).astype(dtype) + 1j * np.random.random( - (100) + y = np.random.random(100).astype(dtype) + 1j * np.random.random( + 100 ).astype(dtype) np_result = np.matmul(x, y) diff --git a/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py b/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py index 31141142ccf..b210ffb5d68 100644 --- a/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py +++ b/python/paddle/fluid/tests/unittests/test_crop_tensor_op.py @@ -147,7 +147,7 @@ class TestCropTensorOpTensorAttr(OpTest): shape_tensor = [] for index, ele in enumerate(self.crop_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { 'X': np.random.random(self.x_shape).astype("float64"), @@ -159,7 +159,7 @@ class TestCropTensorOpTensorAttr(OpTest): offsets_tensor = [] for index, ele in enumerate(self.offsets): offsets_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { 'X': np.random.random(self.x_shape).astype("float64"), diff --git a/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py b/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py index 9bf2e65a6f4..6070f5622af 100644 --- a/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py +++ b/python/paddle/fluid/tests/unittests/test_cross_entropy_loss.py @@ -860,7 +860,7 @@ class CrossEntropyLoss(unittest.TestCase): N = 100 C = 200 input_np = np.random.random([N, C]).astype(self.dtype) - label_np = -np.ones((N)).astype(np.int64) + label_np = -np.ones(N).astype(np.int64) paddle.enable_static() prog = fluid.Program() startup_prog = fluid.Program() diff --git a/python/paddle/fluid/tests/unittests/test_cumsum_op.py b/python/paddle/fluid/tests/unittests/test_cumsum_op.py index a56b0f483bc..541387159cf 100644 --- a/python/paddle/fluid/tests/unittests/test_cumsum_op.py +++ b/python/paddle/fluid/tests/unittests/test_cumsum_op.py @@ -186,7 +186,7 @@ class TestSumOp6(TestSumOp1): class TestSumOp7(TestSumOp1): def set_attrs_input_output(self): - self.x = np.random.random((100)).astype(self.dtype_) + self.x = np.random.random(100).astype(self.dtype_) self.out = self.x.cumsum(axis=0) diff --git a/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py b/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py index 0f0ecf3735d..b443743a351 100644 --- a/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py +++ b/python/paddle/fluid/tests/unittests/test_cumulative_trapezoid.py @@ -73,7 +73,7 @@ class TestCumulativeTrapezoidWithOutDxX(TestCumulativeTrapezoidAPI): class TestCumulativeTrapezoidBroadcast(TestCumulativeTrapezoidAPI): def set_args(self): self.y = np.random.random((3, 3, 4)).astype('float32') - self.x = np.random.random((3)).astype('float32') + self.x = np.random.random(3).astype('float32') self.dx = None self.axis = 1 diff --git a/python/paddle/fluid/tests/unittests/test_dropout_op.py b/python/paddle/fluid/tests/unittests/test_dropout_op.py index 3644801f608..6c5d88155bc 100644 --- a/python/paddle/fluid/tests/unittests/test_dropout_op.py +++ b/python/paddle/fluid/tests/unittests/test_dropout_op.py @@ -96,7 +96,7 @@ class TestDropoutOpInput1d(OpTest): self.attrs = {'dropout_prob': 0.0, 'fix_seed': True, 'is_test': False} self.outputs = { 'Out': self.inputs['X'], - 'Mask': np.ones((2000)).astype('uint8'), + 'Mask': np.ones(2000).astype('uint8'), } # Because prim op compare res with dygraph # when p = 0 dropout api return x,in dygraph mode x_grad = out_grad, diff --git a/python/paddle/fluid/tests/unittests/test_eig_op.py b/python/paddle/fluid/tests/unittests/test_eig_op.py index 5d3d556a67f..c2861fc8bbe 100644 --- a/python/paddle/fluid/tests/unittests/test_eig_op.py +++ b/python/paddle/fluid/tests/unittests/test_eig_op.py @@ -348,7 +348,7 @@ class TestEigWrongDimsError(unittest.TestCase): def test_error(self): paddle.device.set_device("cpu") paddle.disable_static() - a = np.random.random((3)).astype('float32') + a = np.random.random(3).astype('float32') x = paddle.to_tensor(a) self.assertRaises(ValueError, paddle.linalg.eig, x) diff --git a/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py b/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py index bf21a60521a..2e518a2afb3 100644 --- a/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py +++ b/python/paddle/fluid/tests/unittests/test_elementwise_gradient_op.py @@ -106,7 +106,7 @@ class TestElementWiseAddOp(unittest.TestCase): def test_check_forward_backward_with_scale_and_bias(self): np.random.seed(123) self.x = np.random.random((4, 32, 220, 220)).astype(np.float32) - self.y = np.random.random((32)).astype(np.float32) + self.y = np.random.random(32).astype(np.float32) self.out = self.x + self.y.reshape(1, 32, 1, 1) self.axis = 1 self.check_forward_backward() diff --git a/python/paddle/fluid/tests/unittests/test_empty_op.py b/python/paddle/fluid/tests/unittests/test_empty_op.py index 197568f8305..9f8c2a1eaeb 100644 --- a/python/paddle/fluid/tests/unittests/test_empty_op.py +++ b/python/paddle/fluid/tests/unittests/test_empty_op.py @@ -160,7 +160,7 @@ class TestEmptyOp_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/test_expand_op.py b/python/paddle/fluid/tests/unittests/test_expand_op.py index 420aeff5169..f4d6258d820 100644 --- a/python/paddle/fluid/tests/unittests/test_expand_op.py +++ b/python/paddle/fluid/tests/unittests/test_expand_op.py @@ -87,7 +87,7 @@ class TestExpandOpRank1_tensor_attr(OpTest): expand_times_tensor = [] for index, ele in enumerate(self.expand_times): expand_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_expand_v2_op.py b/python/paddle/fluid/tests/unittests/test_expand_v2_op.py index 3a17a4f1bc8..5c0f6ff707f 100644 --- a/python/paddle/fluid/tests/unittests/test_expand_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_expand_v2_op.py @@ -89,7 +89,7 @@ class TestExpandV2OpRank1_tensor_attr(OpTest): expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_fc_op.py b/python/paddle/fluid/tests/unittests/test_fc_op.py index 47cf8c46a58..44b8e2f599d 100644 --- a/python/paddle/fluid/tests/unittests/test_fc_op.py +++ b/python/paddle/fluid/tests/unittests/test_fc_op.py @@ -51,7 +51,7 @@ class MatrixGenerate: if bias_dims == 2: self.bias = np.random.random((1, oc)).astype("float32") else: - self.bias = np.random.random((oc)).astype("float32") + self.bias = np.random.random(oc).astype("float32") class TestFCOp(OpTest): diff --git a/python/paddle/fluid/tests/unittests/test_fill_constant_op.py b/python/paddle/fluid/tests/unittests/test_fill_constant_op.py index aea337e67c1..8aeb64c1a46 100644 --- a/python/paddle/fluid/tests/unittests/test_fill_constant_op.py +++ b/python/paddle/fluid/tests/unittests/test_fill_constant_op.py @@ -155,7 +155,7 @@ class TestFillConstantOp1_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} @@ -180,7 +180,7 @@ class TestFillConstantOp2_ShapeTensorList(OpTest): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/test_full_like_op.py b/python/paddle/fluid/tests/unittests/test_full_like_op.py index c5717d69931..63bfb7be24f 100644 --- a/python/paddle/fluid/tests/unittests/test_full_like_op.py +++ b/python/paddle/fluid/tests/unittests/test_full_like_op.py @@ -68,7 +68,7 @@ class TestFullOp(unittest.TestCase): paddle.disable_static() input = paddle.arange(6, 10, dtype='float32') out = paddle.full_like(input, fill_value=888.88, dtype='float32') - out_numpy = np.random.random((4)).astype("float32") + out_numpy = np.random.random(4).astype("float32") out_numpy.fill(888.88) self.assertTrue((out.numpy() == out_numpy).all(), True) paddle.enable_static() @@ -77,7 +77,7 @@ class TestFullOp(unittest.TestCase): paddle.disable_static() input = paddle.arange(6, 10, dtype='float32') out = paddle.full_like(input, fill_value=float('inf')) - out_numpy = np.random.random((4)).astype("float32") + out_numpy = np.random.random(4).astype("float32") out_numpy.fill(float('inf')) self.assertTrue((out.numpy() == out_numpy).all(), True) paddle.enable_static() diff --git a/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py b/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py index aa49b29916e..c8739c38e76 100644 --- a/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py +++ b/python/paddle/fluid/tests/unittests/test_fused_feedforward_op.py @@ -275,16 +275,16 @@ class APITestStaticFusedFFN(unittest.TestCase): linear1_weight_data = np.random.random( (d_model, dim_feedforward) ).astype(dtype) - linear1_bias_data = np.zeros((dim_feedforward)).astype(dtype) + linear1_bias_data = np.zeros(dim_feedforward).astype(dtype) linear2_weight_data = np.random.random( (dim_feedforward, d_model) ).astype(dtype) - linear2_bias_data = np.zeros((d_model)).astype(dtype) + linear2_bias_data = np.zeros(d_model).astype(dtype) - ln1_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln1_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) - ln2_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln2_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) + ln1_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln1_bias_data = np.zeros(d_model).astype(layer_norm_dtype) + ln2_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln2_bias_data = np.zeros(d_model).astype(layer_norm_dtype) res_list = [fused_out, ln_out] real_res = [] diff --git a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py index 95d74544e67..9830fe455b8 100644 --- a/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py +++ b/python/paddle/fluid/tests/unittests/test_fused_gate_attention_op.py @@ -100,7 +100,7 @@ class TestFusedGateAttentionOp(OpTest): self.gating_b = _random((self.num_heads, self.head_dim)) self.output_w = _random((self.num_heads, self.head_dim, self.out_dim)) - self.output_b = _random((self.out_dim)) + self.output_b = _random(self.out_dim) self.dout = _random( (self.batch_size, self.msa_len, self.res_len, self.q_dim) diff --git a/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py b/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py index 1cb9a9c9172..3b8bf23e600 100644 --- a/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_gaussian_random_op.py @@ -123,7 +123,7 @@ class TestGaussianRandomOp_ShapeTensorList(TestGaussianRandomOp): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.attrs = { diff --git a/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py b/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py index 2d7e6508661..98234e1613f 100644 --- a/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_gru_rnn_op.py @@ -136,7 +136,7 @@ class TestGRUOp(OpTest): (self.num_layers * self.direction_num, batch_size, self.hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -162,7 +162,7 @@ class TestGRUOp(OpTest): self.outputs = { 'Out': output, 'State': [('last_hidden', last_hidden)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py b/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py index 5525854b008..c0396a7a029 100644 --- a/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py +++ b/python/paddle/fluid/tests/unittests/test_gumbel_softmax_op.py @@ -186,7 +186,7 @@ class TestGumbelSoftmaxOpSampleDistribution(OpTest): # Construct statistics z for samples and # z is approximately N(0,1) for unbiased count expected = self.probs * self.shape[0] - z = (self.counts - expected) / np.sqrt((expected * (1 - self.probs))) + z = (self.counts - expected) / np.sqrt(expected * (1 - self.probs)) # A (lazy) approximate 99% two-sided test: # occurs with prob alpha~>=0.01 if unbiased self.assertLess(np.max(np.abs(z)).item(), 2.58) diff --git a/python/paddle/fluid/tests/unittests/test_imperative_framework.py b/python/paddle/fluid/tests/unittests/test_imperative_framework.py index 8f4f73daa0a..3c81da2c6d5 100644 --- a/python/paddle/fluid/tests/unittests/test_imperative_framework.py +++ b/python/paddle/fluid/tests/unittests/test_imperative_framework.py @@ -64,7 +64,7 @@ class TestDygraphFramework(unittest.TestCase): "backward should not be usable in static graph mode" ) except AssertionError as e: - self.assertTrue((e is not None)) + self.assertTrue(e is not None) def test_dygraph_to_string(self): np_inp = np.array([[1.0, 2.0], [3.0, 4.0]], dtype=np.float32) diff --git a/python/paddle/fluid/tests/unittests/test_initializer.py b/python/paddle/fluid/tests/unittests/test_initializer.py index b87444e19dd..79b4782e6b6 100644 --- a/python/paddle/fluid/tests/unittests/test_initializer.py +++ b/python/paddle/fluid/tests/unittests/test_initializer.py @@ -38,7 +38,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-1, 1)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -632,7 +632,7 @@ class TestNumpyArrayInitializer(unittest.TestCase): program = framework.Program() block = program.global_block() - np_array = numpy.random.random((10000)).astype(dtype) + np_array = numpy.random.random(10000).astype(dtype) for _ in range(2): block.create_parameter( dtype=np_array.dtype, diff --git a/python/paddle/fluid/tests/unittests/test_initializer_nn.py b/python/paddle/fluid/tests/unittests/test_initializer_nn.py index b590fbe5212..f90902631c5 100644 --- a/python/paddle/fluid/tests/unittests/test_initializer_nn.py +++ b/python/paddle/fluid/tests/unittests/test_initializer_nn.py @@ -644,7 +644,7 @@ class TestAssign(unittest.TestCase): program = framework.Program() block = program.global_block() - np_array = numpy.random.random((10000)).astype(dtype) + np_array = numpy.random.random(10000).astype(dtype) for _ in range(2): block.create_parameter( dtype=np_array.dtype, diff --git a/python/paddle/fluid/tests/unittests/test_jit_save_load.py b/python/paddle/fluid/tests/unittests/test_jit_save_load.py index e94c44193e7..d04357e201c 100644 --- a/python/paddle/fluid/tests/unittests/test_jit_save_load.py +++ b/python/paddle/fluid/tests/unittests/test_jit_save_load.py @@ -1102,7 +1102,7 @@ class TestJitSaveLoadEmptyLayer(unittest.TestCase): def test_save_load_empty_layer(self): layer = EmptyLayer() - x = paddle.to_tensor(np.random.random((10)).astype('float32')) + x = paddle.to_tensor(np.random.random(10).astype('float32')) out = layer(x) paddle.jit.save(layer, self.model_path) load_layer = paddle.jit.load(self.model_path) @@ -1124,8 +1124,8 @@ class TestJitSaveLoadNoParamLayer(unittest.TestCase): def test_save_load_no_param_layer(self): layer = NoParamLayer() - x = paddle.to_tensor(np.random.random((5)).astype('float32')) - y = paddle.to_tensor(np.random.random((5)).astype('float32')) + x = paddle.to_tensor(np.random.random(5).astype('float32')) + y = paddle.to_tensor(np.random.random(5).astype('float32')) out = layer(x, y) paddle.jit.save(layer, self.model_path) load_layer = paddle.jit.load(self.model_path) @@ -1433,7 +1433,7 @@ class TestJitSaveLoadFinetuneLoad(unittest.TestCase): result_11 = layer_finetune(inps1) self.assertTrue(float((result_00 - result_10).abs().max()) < 1e-5) - self.assertTrue(float(((result_01 - result_11)).abs().max()) < 1e-5) + self.assertTrue(float((result_01 - result_11).abs().max()) < 1e-5) # NOTE(weixin): When there are multiple test functions in an diff --git a/python/paddle/fluid/tests/unittests/test_linear.py b/python/paddle/fluid/tests/unittests/test_linear.py index 4d4cdde6ede..8823a4f40be 100644 --- a/python/paddle/fluid/tests/unittests/test_linear.py +++ b/python/paddle/fluid/tests/unittests/test_linear.py @@ -27,7 +27,7 @@ class LinearTestCase(unittest.TestCase): self.dtype = 'float32' self.input = np.ones((3, 1, 2)).astype(self.dtype) self.weight = np.ones((2, 2)).astype(self.dtype) - self.bias = np.ones((2)).astype(self.dtype) + self.bias = np.ones(2).astype(self.dtype) self.place = ( paddle.CUDAPlace(0) if core.is_compiled_with_cuda() diff --git a/python/paddle/fluid/tests/unittests/test_linear_interp_op.py b/python/paddle/fluid/tests/unittests/test_linear_interp_op.py index db61b762c89..e18dd205794 100755 --- a/python/paddle/fluid/tests/unittests/test_linear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_linear_interp_op.py @@ -227,7 +227,7 @@ class TestLinearInterpOpSizeTensor(TestLinearInterpOp): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py index 323e1d8ca5b..4bb0a210c66 100755 --- a/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_linear_interp_v2_op.py @@ -295,7 +295,7 @@ class TestLinearInterpOpSizeTensor(TestLinearInterpOp): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py b/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py index d1b944b29c3..f81eb531424 100644 --- a/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_lstm_cudnn_op.py @@ -474,7 +474,7 @@ class TestCUDNNLstmOp(OpTest): init_c = np.zeros((self.num_layers, batch_size, hidden_size)).astype( self.dtype ) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") if core.is_compiled_with_rocm(): for i in range(len(flat_w)): @@ -508,7 +508,7 @@ class TestCUDNNLstmOp(OpTest): 'Out': output, "LastH": last_hidden, 'LastC': last_cell, - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'StateOut': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_meshgrid_op.py b/python/paddle/fluid/tests/unittests/test_meshgrid_op.py index 2946008b0fd..cc6f04eb7a9 100644 --- a/python/paddle/fluid/tests/unittests/test_meshgrid_op.py +++ b/python/paddle/fluid/tests/unittests/test_meshgrid_op.py @@ -268,7 +268,7 @@ class TestMeshGrid_ZeroDim(TestMeshgridOp): self.shape = self.get_x_shape() ins = [] outs = [] - ins.append(np.random.random(([])).astype(self.dtype)) + ins.append(np.random.random([]).astype(self.dtype)) ins.append(np.random.random([2]).astype(self.dtype)) ins.append(np.random.random([3]).astype(self.dtype)) for i in range(len(self.shape)): diff --git a/python/paddle/fluid/tests/unittests/test_multi_dot_op.py b/python/paddle/fluid/tests/unittests/test_multi_dot_op.py index 0ac38b9ac08..a5555657c7f 100644 --- a/python/paddle/fluid/tests/unittests/test_multi_dot_op.py +++ b/python/paddle/fluid/tests/unittests/test_multi_dot_op.py @@ -104,7 +104,7 @@ class TestMultiDotOp4Mat(TestMultiDotOp): class TestMultiDotOpFirst1D(TestMultiDotOp): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -112,7 +112,7 @@ class TestMultiDotOpFirst1D(TestMultiDotOp): class TestMultiDotOp3MatFirst1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.C = np.random.random((3, 3)).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} @@ -121,7 +121,7 @@ class TestMultiDotOp3MatFirst1D(TestMultiDotOp3Mat): class TestMultiDotOp4MatFirst1D(TestMultiDotOp4Mat): def get_inputs_and_outputs(self): - self.A = np.random.random((4)).astype(self.dtype) + self.A = np.random.random(4).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) self.C = np.random.random((3, 4)).astype(self.dtype) self.D = np.random.random((4, 5)).astype(self.dtype) @@ -139,7 +139,7 @@ class TestMultiDotOp4MatFirst1D(TestMultiDotOp4Mat): class TestMultiDotOpLast1D(TestMultiDotOp): def get_inputs_and_outputs(self): self.A = np.random.random((3, 6)).astype(self.dtype) - self.B = np.random.random((6)).astype(self.dtype) + self.B = np.random.random(6).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -148,7 +148,7 @@ class TestMultiDotOp3MatLast1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): self.A = np.random.random((2, 4)).astype(self.dtype) self.B = np.random.random((4, 3)).astype(self.dtype) - self.C = np.random.random((3)).astype(self.dtype) + self.C = np.random.random(3).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} self.outputs = {'Out': multi_dot([self.A, self.B, self.C])} @@ -163,7 +163,7 @@ class TestMultiDotOp4MatLast1D(TestMultiDotOp4Mat): self.A = np.random.random((2, 3)).astype(self.dtype) self.B = np.random.random((3, 2)).astype(self.dtype) self.C = np.random.random((2, 3)).astype(self.dtype) - self.D = np.random.random((3)).astype(self.dtype) + self.D = np.random.random(3).astype(self.dtype) self.inputs = { 'X': [ ('x0', self.A), @@ -178,7 +178,7 @@ class TestMultiDotOp4MatLast1D(TestMultiDotOp4Mat): class TestMultiDotOpFirstAndLast1D(TestMultiDotOp): def get_inputs_and_outputs(self): self.A = np.random.random((4,)).astype(self.dtype) - self.B = np.random.random((4)).astype(self.dtype) + self.B = np.random.random(4).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B)]} self.outputs = {'Out': multi_dot([self.A, self.B])} @@ -187,7 +187,7 @@ class TestMultiDotOp3MatFirstAndLast1D(TestMultiDotOp3Mat): def get_inputs_and_outputs(self): self.A = np.random.random((6,)).astype(self.dtype) self.B = np.random.random((6, 4)).astype(self.dtype) - self.C = np.random.random((4)).astype(self.dtype) + self.C = np.random.random(4).astype(self.dtype) self.inputs = {'X': [('x0', self.A), ('x1', self.B), ('x2', self.C)]} self.outputs = {'Out': multi_dot([self.A, self.B, self.C])} @@ -197,7 +197,7 @@ class TestMultiDotOp4MatFirstAndLast1D(TestMultiDotOp4Mat): self.A = np.random.random((3,)).astype(self.dtype) self.B = np.random.random((3, 4)).astype(self.dtype) self.C = np.random.random((4, 2)).astype(self.dtype) - self.D = np.random.random((2)).astype(self.dtype) + self.D = np.random.random(2).astype(self.dtype) self.inputs = { 'X': [ ('x0', self.A), diff --git a/python/paddle/fluid/tests/unittests/test_mv_op.py b/python/paddle/fluid/tests/unittests/test_mv_op.py index 93b541e7ecc..9f0840fb123 100644 --- a/python/paddle/fluid/tests/unittests/test_mv_op.py +++ b/python/paddle/fluid/tests/unittests/test_mv_op.py @@ -37,7 +37,7 @@ class TestMVOp(OpTest): def init_config(self): self.x = np.random.random((2, 100)).astype("float64") - self.vec = np.random.random((100)).astype("float64") + self.vec = np.random.random(100).astype("float64") class TestMVAPI(unittest.TestCase): @@ -46,7 +46,7 @@ class TestMVAPI(unittest.TestCase): self.x_data = np.random.random((5, 100)).astype("float64") self.x = paddle.to_tensor(self.x_data) - self.vec_data = np.random.random((100)).astype("float64") + self.vec_data = np.random.random(100).astype("float64") self.vec = paddle.to_tensor(self.vec_data) z = paddle.mv(self.x, self.vec) np_z = z.numpy() diff --git a/python/paddle/fluid/tests/unittests/test_nanmedian.py b/python/paddle/fluid/tests/unittests/test_nanmedian.py index 42509cdb5f2..88950271d32 100644 --- a/python/paddle/fluid/tests/unittests/test_nanmedian.py +++ b/python/paddle/fluid/tests/unittests/test_nanmedian.py @@ -189,7 +189,7 @@ class TestNanmedian(unittest.TestCase): x_np[2, 3:] = np.nan x_np_sorted = np.sort(x_np) nan_counts = np.count_nonzero(np.isnan(x_np).astype(np.int32), axis=1) - np_grad = np.zeros((shape)) + np_grad = np.zeros(shape) for i in range(shape[0]): valid_cnts = shape[1] - nan_counts[i] if valid_cnts == 0: diff --git a/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py b/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py index 86233f7411a..cb153bdc55b 100755 --- a/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_nearest_interp_op.py @@ -374,7 +374,7 @@ class TestNearestInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py index 2447818c896..6ddbdea0ece 100755 --- a/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_nearest_interp_v2_op.py @@ -874,7 +874,7 @@ class TestNearestInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_npair_loss_op.py b/python/paddle/fluid/tests/unittests/test_npair_loss_op.py index a4a1b16efc6..b49cdf0547c 100755 --- a/python/paddle/fluid/tests/unittests/test_npair_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_npair_loss_op.py @@ -128,7 +128,7 @@ class TestNpairLossOpError(unittest.TestCase): with program_guard(Program(), Program()): anchor_np = np.random.random((2, 4)).astype("float32") positive_np = np.random.random((2, 4)).astype("float32") - labels_np = np.random.random((2)).astype("float32") + labels_np = np.random.random(2).astype("float32") anchor_data = paddle.static.data( name='anchor', shape=[2, 4], dtype='float32' ) diff --git a/python/paddle/fluid/tests/unittests/test_numel_op.py b/python/paddle/fluid/tests/unittests/test_numel_op.py index e4122bc0fb1..81ddbdc45a0 100644 --- a/python/paddle/fluid/tests/unittests/test_numel_op.py +++ b/python/paddle/fluid/tests/unittests/test_numel_op.py @@ -27,7 +27,7 @@ class TestNumelOp(OpTest): self.op_type = "size" self.python_api = paddle.numel self.init() - x = np.random.random((self.shape)).astype(self.dtype) + x = np.random.random(self.shape).astype(self.dtype) self.inputs = { 'Input': x, } @@ -82,7 +82,7 @@ class TestNumelOpBF16(OpTest): self.python_api = paddle.numel self.dtype = np.uint16 self.init() - x = np.random.random((self.shape)).astype(np.float32) + x = np.random.random(self.shape).astype(np.float32) self.inputs = {'Input': convert_float_to_uint16(x)} self.outputs = {'Out': np.array([np.size(x)])} diff --git a/python/paddle/fluid/tests/unittests/test_pool2d_op.py b/python/paddle/fluid/tests/unittests/test_pool2d_op.py index f4cf8943ea3..7a835f0fb44 100644 --- a/python/paddle/fluid/tests/unittests/test_pool2d_op.py +++ b/python/paddle/fluid/tests/unittests/test_pool2d_op.py @@ -778,7 +778,7 @@ class TestCase4_AsyPadding(TestCase4): self.shape = [2, 3, 7, 7] -class TestCase5_AsyPadding((TestCase5)): +class TestCase5_AsyPadding(TestCase5): def init_test_case(self): self.ksize = [3, 3] self.strides = [1, 1] diff --git a/python/paddle/fluid/tests/unittests/test_pylayer_op.py b/python/paddle/fluid/tests/unittests/test_pylayer_op.py index b270a8b0d7f..ce036905138 100644 --- a/python/paddle/fluid/tests/unittests/test_pylayer_op.py +++ b/python/paddle/fluid/tests/unittests/test_pylayer_op.py @@ -55,7 +55,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_return_none_with_no_grad(self): @@ -91,7 +91,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_single_output(self): @@ -119,7 +119,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_simple_pylayer_multi_output(self): @@ -149,7 +149,7 @@ class TestPyLayer(unittest.TestCase): z2.mean().backward() self.assertTrue( - np.max(np.abs((input1.grad.numpy() - input2.grad.numpy()))) < 1e-10 + np.max(np.abs(input1.grad.numpy() - input2.grad.numpy())) < 1e-10 ) def test_pylayer_num_output_match(self): diff --git a/python/paddle/fluid/tests/unittests/test_randint_op.py b/python/paddle/fluid/tests/unittests/test_randint_op.py index 678cb3a5a2e..073b8255be9 100644 --- a/python/paddle/fluid/tests/unittests/test_randint_op.py +++ b/python/paddle/fluid/tests/unittests/test_randint_op.py @@ -29,7 +29,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-10, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -74,7 +74,7 @@ class TestRandintOp_attr_tensorlist(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -160,7 +160,7 @@ class TestRandintImperative(unittest.TestCase): x3 = paddle.tensor.random.randint(n) for i in [x1, x2, x3]: for j in i.numpy().tolist(): - self.assertTrue((j >= 0 and j < n)) + self.assertTrue(j >= 0 and j < n) paddle.enable_static() diff --git a/python/paddle/fluid/tests/unittests/test_randperm_op.py b/python/paddle/fluid/tests/unittests/test_randperm_op.py index b482a0666aa..eaecf087f9f 100644 --- a/python/paddle/fluid/tests/unittests/test_randperm_op.py +++ b/python/paddle/fluid/tests/unittests/test_randperm_op.py @@ -63,7 +63,7 @@ class TestRandpermOp(OpTest): self.dtype = "int64" self.inputs = {} - self.outputs = {"Out": np.zeros((self.n)).astype(self.dtype)} + self.outputs = {"Out": np.zeros(self.n).astype(self.dtype)} self.init_attrs() self.attrs = { "n": self.n, diff --git a/python/paddle/fluid/tests/unittests/test_rank_attention_op.py b/python/paddle/fluid/tests/unittests/test_rank_attention_op.py index 8428e6a043e..d699a6311a3 100644 --- a/python/paddle/fluid/tests/unittests/test_rank_attention_op.py +++ b/python/paddle/fluid/tests/unittests/test_rank_attention_op.py @@ -24,7 +24,7 @@ from paddle.fluid import core def gen_input_help(input, rank_offset, max_rank, max_size): input_row, input_col = input.shape max_ins = np.max((max_size, input_row)) - input_help = np.zeros((max_ins * max_rank * input_col)) + input_help = np.zeros(max_ins * max_rank * input_col) ins_rank = np.zeros((max_ins, 1)) ins_rank.fill(-1) diff --git a/python/paddle/fluid/tests/unittests/test_reshape_op.py b/python/paddle/fluid/tests/unittests/test_reshape_op.py index 0647cde61cc..24b8661694e 100755 --- a/python/paddle/fluid/tests/unittests/test_reshape_op.py +++ b/python/paddle/fluid/tests/unittests/test_reshape_op.py @@ -174,7 +174,7 @@ class TestReshapeOp_attr_ShapeTensor(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py b/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py index bf01589c577..a7104e37399 100644 --- a/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py +++ b/python/paddle/fluid/tests/unittests/test_rnn_cell_api.py @@ -120,7 +120,7 @@ class TestRnnError(unittest.TestCase): self.assertRaises(TypeError, test_initial_states_type) def test_sequence_length_type(): - np_sequence_length = np.random.random((batch_size)).astype( + np_sequence_length = np.random.random(batch_size).astype( "float32" ) dynamic_rnn( diff --git a/python/paddle/fluid/tests/unittests/test_rnn_op.py b/python/paddle/fluid/tests/unittests/test_rnn_op.py index f4fc13996fd..73ef3381653 100644 --- a/python/paddle/fluid/tests/unittests/test_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_rnn_op.py @@ -137,7 +137,7 @@ class TestRNNOp(OpTest): init_c = np.zeros( (self.num_layers * self.direction_num, batch_size, hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -163,7 +163,7 @@ class TestRNNOp(OpTest): self.outputs = { 'Out': output, "State": [('last_hidden', last_hidden), ('last_cell', last_cell)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_seed_op.py b/python/paddle/fluid/tests/unittests/test_seed_op.py index d72bc8601d7..426dab42148 100644 --- a/python/paddle/fluid/tests/unittests/test_seed_op.py +++ b/python/paddle/fluid/tests/unittests/test_seed_op.py @@ -28,7 +28,7 @@ class TestSeedOpFixSeed(OpTest): self.op_type = "seed" self.inputs = {} self.attrs = {"seed": 123} - self.outputs = {"Out": np.asarray((123)).astype('int')} + self.outputs = {"Out": np.asarray(123).astype('int')} def test_check_output(self): self.check_output() @@ -39,7 +39,7 @@ class TestSeedOpDiffSeed(OpTest): self.op_type = "seed" self.inputs = {} self.attrs = {"seed": 0} - self.outputs = {"Out": np.asarray((123)).astype('int')} + self.outputs = {"Out": np.asarray(123).astype('int')} def test_check_output(self): self.check_output(no_check_set=["Out"]) diff --git a/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py b/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py index dfeeef8ad29..abf4ba677ba 100644 --- a/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_sigmoid_focal_loss_op.py @@ -36,8 +36,8 @@ def sigmoid_focal_loss_forward( a = int(idx / num_classes) d = int(idx % num_classes) label = label_data[a] - c_pos = float((int(label) == int(d + 1))) - c_neg = float(((int(label) != -1) & (int(label) != (d + 1)))) + c_pos = float(int(label) == int(d + 1)) + c_neg = float((int(label) != -1) & (int(label) != (d + 1))) fg_num = max(fg_num_data, 1) z_neg = (1.0 - alpha) / fg_num z_pos = alpha / fg_num diff --git a/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py b/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py index 4fe7f5a9dd0..b0901005204 100644 --- a/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py +++ b/python/paddle/fluid/tests/unittests/test_simple_rnn_op.py @@ -128,7 +128,7 @@ class TestSimpleRNNOp(OpTest): (self.num_layers * self.direction_num, batch_size, hidden_size) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -154,7 +154,7 @@ class TestSimpleRNNOp(OpTest): self.outputs = { 'Out': output, 'State': [('last_hidden', last_hidden)], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/test_slice_op.py b/python/paddle/fluid/tests/unittests/test_slice_op.py index 4e8e62b1f44..0314a37170d 100644 --- a/python/paddle/fluid/tests/unittests/test_slice_op.py +++ b/python/paddle/fluid/tests/unittests/test_slice_op.py @@ -172,7 +172,7 @@ class TestSliceOp_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int64') * ele) + ("x" + str(index), np.ones(1).astype('int64') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -212,7 +212,7 @@ class TestSliceOp_decs_dim_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -379,7 +379,7 @@ class TestSliceOp_starts_OneTensor_ends_ListTensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("y" + str(index), np.ones((1)).astype('int32') * ele) + ("y" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_split_op.py b/python/paddle/fluid/tests/unittests/test_split_op.py index 7d50d9586b3..f0929e5758d 100644 --- a/python/paddle/fluid/tests/unittests/test_split_op.py +++ b/python/paddle/fluid/tests/unittests/test_split_op.py @@ -154,7 +154,7 @@ class TestSplitOp_SectionsTensor(OpTest): sections_tensor = [] for index, ele in enumerate(self.sections): sections_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SectionsTensorList'] = sections_tensor diff --git a/python/paddle/fluid/tests/unittests/test_strided_slice_op.py b/python/paddle/fluid/tests/unittests/test_strided_slice_op.py index fb841244da1..dc5397fb4f4 100644 --- a/python/paddle/fluid/tests/unittests/test_strided_slice_op.py +++ b/python/paddle/fluid/tests/unittests/test_strided_slice_op.py @@ -324,7 +324,7 @@ class TestStridedSliceOp_starts_ListTensor(OpTest): starts_tensor = [] for index, ele in enumerate(self.starts): starts_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'StartsTensorList': starts_tensor} @@ -366,7 +366,7 @@ class TestStridedSliceOp_ends_ListTensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {'Input': self.input, 'EndsTensorList': ends_tensor} @@ -477,7 +477,7 @@ class TestStridedSliceOp_listTensor_Tensor(OpTest): ends_tensor = [] for index, ele in enumerate(self.ends): ends_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.op_type = "strided_slice" self.python_api = paddle.strided_slice diff --git a/python/paddle/fluid/tests/unittests/test_tile_op.py b/python/paddle/fluid/tests/unittests/test_tile_op.py index ec450cd888e..61901ce1df4 100644 --- a/python/paddle/fluid/tests/unittests/test_tile_op.py +++ b/python/paddle/fluid/tests/unittests/test_tile_op.py @@ -111,7 +111,7 @@ class TestTileOpRank1_tensor_attr(OpTest): repeat_times_tensor = [] for index, ele in enumerate(self.repeat_times): repeat_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_trapezoid.py b/python/paddle/fluid/tests/unittests/test_trapezoid.py index ed5bbc2ae21..f622c6b7466 100644 --- a/python/paddle/fluid/tests/unittests/test_trapezoid.py +++ b/python/paddle/fluid/tests/unittests/test_trapezoid.py @@ -148,7 +148,7 @@ class TestTrapezoidWithOutDxX(TestTrapezoidAPI): class TestTrapezoidBroadcast(TestTrapezoidAPI): def set_args(self): self.y = np.random.random((3, 3, 4)).astype('float32') - self.x = np.random.random((3)).astype('float32') + self.x = np.random.random(3).astype('float32') self.dx = None self.axis = 1 diff --git a/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py b/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py index 8001997696e..031639fc080 100755 --- a/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py +++ b/python/paddle/fluid/tests/unittests/test_trilinear_interp_op.py @@ -528,7 +528,7 @@ class TestTrilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py b/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py index be13d5ebda1..6b774fafe3f 100755 --- a/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py +++ b/python/paddle/fluid/tests/unittests/test_trilinear_interp_v2_op.py @@ -888,7 +888,7 @@ class TestTrilinearInterpOp_attr_tensor(OpTest): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py b/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py index 0e08c3d8bc0..e76bbc80033 100644 --- a/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py +++ b/python/paddle/fluid/tests/unittests/test_uniform_random_bf16_op.py @@ -70,7 +70,7 @@ class TestUniformRandomOpBF16AttrTensorList(TestUniformRandomOpBF16): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() diff --git a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py index 3f550e3ed7c..de1984098b9 100644 --- a/python/paddle/fluid/tests/unittests/test_uniform_random_op.py +++ b/python/paddle/fluid/tests/unittests/test_uniform_random_op.py @@ -33,7 +33,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-5, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -46,7 +46,7 @@ def output_hist_diag(out): hist, _ = np.histogram(out, range=(-5, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob @@ -58,7 +58,7 @@ class TestUniformRandomOp_attr_tensorlist(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int64") * ele) + ("x" + str(index), np.ones(1).astype("int64") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -90,7 +90,7 @@ class TestUniformRandomOp_attr_tensorlist_int32(OpTest): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype("int32") * ele) + ("x" + str(index), np.ones(1).astype("int32") * ele) ) self.inputs = {'ShapeTensorList': shape_tensor} self.init_attrs() @@ -471,7 +471,7 @@ class TestUniformRandomDygraphMode(unittest.TestCase): x = paddle.uniform([10], dtype="float32", min=0.0, max=1.0) x_np = x.numpy() for i in range(10): - self.assertTrue((x_np[i] > 0 and x_np[i] < 1.0)) + self.assertTrue(x_np[i] > 0 and x_np[i] < 1.0) class TestUniformRandomBatchSizeLikeOpError(unittest.TestCase): @@ -562,7 +562,7 @@ class TestUniformDygraphMode(unittest.TestCase): ) x_np = x.numpy() for i in range(10): - self.assertTrue((x_np[i] > 0 and x_np[i] < 1.0)) + self.assertTrue(x_np[i] > 0 and x_np[i] < 1.0) class TestUniformDtype(unittest.TestCase): diff --git a/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py b/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py index ea5044d1f81..b7b4c185e97 100755 --- a/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py +++ b/python/paddle/fluid/tests/unittests/test_unsqueeze2_op.py @@ -120,7 +120,7 @@ class TestUnsqueezeOp_AxesTensorList(OpTest): axes_tensor_list = [] for index, ele in enumerate(self.axes): axes_tensor_list.append( - ("axes" + str(index), np.ones((1)).astype('int32') * ele) + ("axes" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/test_variable.py b/python/paddle/fluid/tests/unittests/test_variable.py index 8e5752d1985..c26053a7342 100644 --- a/python/paddle/fluid/tests/unittests/test_variable.py +++ b/python/paddle/fluid/tests/unittests/test_variable.py @@ -173,7 +173,7 @@ class TestVariable(unittest.TestCase): y_1 = y[:, 0] feeder = fluid.DataFeeder(place=place, feed_list=[x]) data = [] - data.append((np.random.randint(10, size=[13]).astype('float32'))) + data.append(np.random.randint(10, size=[13]).astype('float32')) exe.run(fluid.default_startup_program()) local_out = exe.run( diff --git a/python/paddle/fluid/tests/unittests/test_yolo_box_op.py b/python/paddle/fluid/tests/unittests/test_yolo_box_op.py index a572b91b0b5..81445c081fe 100644 --- a/python/paddle/fluid/tests/unittests/test_yolo_box_op.py +++ b/python/paddle/fluid/tests/unittests/test_yolo_box_op.py @@ -21,13 +21,13 @@ import paddle def sigmoid(x): - return 1.0 / (1.0 + np.exp(((-1.0) * x))) + return 1.0 / (1.0 + np.exp((-1.0) * x)) def YoloBox(x, img_size, attrs): (n, c, h, w) = x.shape anchors = attrs['anchors'] - an_num = int((len(anchors) // 2)) + an_num = int(len(anchors) // 2) class_num = attrs['class_num'] conf_thresh = attrs['conf_thresh'] downsample = attrs['downsample_ratio'] @@ -145,7 +145,7 @@ class TestYoloBoxOp(OpTest): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -166,7 +166,7 @@ class TestYoloBoxOp(OpTest): class TestYoloBoxOpNoClipBbox(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -187,7 +187,7 @@ class TestYoloBoxOpNoClipBbox(TestYoloBoxOp): class TestYoloBoxOpScaleXY(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -208,7 +208,7 @@ class TestYoloBoxOpScaleXY(TestYoloBoxOp): class TestYoloBoxOpIoUAware(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 @@ -295,7 +295,7 @@ class TestYoloBoxStatic(unittest.TestCase): class TestYoloBoxOpHW(TestYoloBoxOp): def initTestCase(self): self.anchors = [10, 13, 16, 30, 33, 23] - an_num = int((len(self.anchors) // 2)) + an_num = int(len(self.anchors) // 2) self.batch_size = 32 self.class_num = 2 self.conf_thresh = 0.5 diff --git a/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py b/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py index 288ed525a78..79994d87e42 100644 --- a/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py +++ b/python/paddle/fluid/tests/unittests/test_yolov3_loss_op.py @@ -81,7 +81,7 @@ def YOLOv3Loss(x, gtbox, gtlabel, gtscore, attrs): bias_x_y = -0.5 * (scale_x_y - 1.0) input_size = downsample_ratio * h x = x.reshape((n, mask_num, 5 + class_num, h, w)).transpose((0, 1, 3, 4, 2)) - loss = np.zeros((n)).astype('float64') + loss = np.zeros(n).astype('float64') smooth_weight = min(1.0 / class_num, 1.0 / 40) label_pos = 1.0 - smooth_weight if use_label_smooth else 1.0 diff --git a/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py index 9645fbb7d00..b071db95b40 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_activation_op_xpu.py @@ -1011,7 +1011,7 @@ class XPUTestSoftReluOP(XPUOpTestWrapper): t = np.copy(x) t[t < -threshold] = -threshold t[t > threshold] = threshold - out = np.log((np.exp(t) + 1)) + out = np.log(np.exp(t) + 1) self.inputs = {'X': x} self.outputs = {'Out': out} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py index 0899d6f45e1..d30fdbed09d 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_adamw_op_xpu.py @@ -484,30 +484,22 @@ class XPUTestAdamwOp2(XPUOpTestWrapper): out = linear1(x) out = linear2(out) - fc1_w_mon1 = np.zeros((linear1.weight.shape)).astype( + fc1_w_mon1 = np.zeros(linear1.weight.shape).astype( "float32" ) - fc1_w_mon2 = np.zeros((linear1.weight.shape)).astype( + fc1_w_mon2 = np.zeros(linear1.weight.shape).astype( "float32" ) - fc1_b_mon1 = np.zeros((linear1.bias.shape)).astype( + fc1_b_mon1 = np.zeros(linear1.bias.shape).astype("float32") + fc1_b_mon2 = np.zeros(linear1.bias.shape).astype("float32") + fc2_w_mon1 = np.zeros(linear2.weight.shape).astype( "float32" ) - fc1_b_mon2 = np.zeros((linear1.bias.shape)).astype( - "float32" - ) - fc2_w_mon1 = np.zeros((linear2.weight.shape)).astype( - "float32" - ) - fc2_w_mon2 = np.zeros((linear2.weight.shape)).astype( - "float32" - ) - fc2_b_mon1 = np.zeros((linear2.bias.shape)).astype( - "float32" - ) - fc2_b_mon2 = np.zeros((linear2.bias.shape)).astype( + fc2_w_mon2 = np.zeros(linear2.weight.shape).astype( "float32" ) + fc2_b_mon1 = np.zeros(linear2.bias.shape).astype("float32") + fc2_b_mon2 = np.zeros(linear2.bias.shape).astype("float32") cost = paddle.nn.functional.square_error_cost( input=out, label=y diff --git a/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py index 388ace1ed33..e171625dd43 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_amp_check_finite_and_scale_op_xpu.py @@ -40,7 +40,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): self.op_type = "check_finite_and_unscale" self.init_dtype() x = np.random.random((8, 8)).astype(self.dtype) - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { 'FoundInfinite': np.array([0]), @@ -64,7 +64,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.nan x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -91,7 +91,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.nan x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) myscale = np.array([0.05]).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { @@ -118,7 +118,7 @@ class XPUTestCheckFiniteAndUnscaleOp(XPUOpTestWrapper): idx2 = np.random.randint(255) x[idx1][idx2] = np.inf x[idx2][idx1] = np.nan - scale = np.random.random((1)).astype(np.float32) + scale = np.random.random(1).astype(np.float32) myscale = np.array([0.05]).astype(self.dtype) self.inputs = {'X': [('x0', x)], 'Scale': scale} self.outputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py index 5067baf57c9..ebd48f55d57 100755 --- a/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_bilinear_interp_v2_op_xpu.py @@ -409,7 +409,7 @@ class XPUTestBilinearInterpV2Op(XPUOpTestWrapper): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py index 7b11ec36675..3f1b1066b91 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_coalesce_tensor_op_xpu.py @@ -96,7 +96,7 @@ class XPUTestCoalesceTensorOp(XPUOpTestWrapper): coalesce_tensor_var = np.concatenate([input for input in inputs]) if set_constant: coalesce_tensor_var = ( - np.ones((len(coalesce_tensor_var))) * constant + np.ones(len(coalesce_tensor_var)) * constant ) outputs = [ ( diff --git a/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py index 0bc1094888b..87241881275 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_empty_op_xpu.py @@ -126,7 +126,7 @@ class XPUTestEmptyOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py index 4f7e9dd5e9f..f7098282a62 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_expand_v2_op_xpu.py @@ -118,7 +118,7 @@ class XPUTestExpandV2Op(XPUOpTestWrapper): expand_shapes_tensor = [] for index, ele in enumerate(self.expand_shape): expand_shapes_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py index 086376f960c..4bd9abae9a5 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_fill_constant_op_xpu.py @@ -147,7 +147,7 @@ class XPUTestFillConstantOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} @@ -178,7 +178,7 @@ class XPUTestFillConstantOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = {"ShapeTensorList": shape_tensor_list} diff --git a/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py index f1aac97dcba..feb7549a33e 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_fused_feedforward_op_xpu.py @@ -265,16 +265,16 @@ class APITestStaticFusedFFN(unittest.TestCase): linear1_weight_data = np.random.random( (d_model, dim_feedforward) ).astype(dtype) - linear1_bias_data = np.zeros((dim_feedforward)).astype(dtype) + linear1_bias_data = np.zeros(dim_feedforward).astype(dtype) linear2_weight_data = np.random.random( (dim_feedforward, d_model) ).astype(dtype) - linear2_bias_data = np.zeros((d_model)).astype(dtype) + linear2_bias_data = np.zeros(d_model).astype(dtype) - ln1_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln1_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) - ln2_scale_data = np.ones((d_model)).astype(layer_norm_dtype) - ln2_bias_data = np.zeros((d_model)).astype(layer_norm_dtype) + ln1_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln1_bias_data = np.zeros(d_model).astype(layer_norm_dtype) + ln2_scale_data = np.ones(d_model).astype(layer_norm_dtype) + ln2_bias_data = np.zeros(d_model).astype(layer_norm_dtype) res_list = [fused_out, ln_out] real_res = [] diff --git a/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py index 3df6d2978c7..9d5b5e747f4 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_gaussian_random_op_xpu.py @@ -94,7 +94,7 @@ class XPUTestGaussianRandomOp(XPUOpTestWrapper): shape_tensor_list = [] for index, ele in enumerate(self.shape): shape_tensor_list.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.attrs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py index 7a9150312c9..35c362b9a9a 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_nearest_interp_v2_op_xpu.py @@ -470,7 +470,7 @@ class XPUNearestInterpOpWrapper(XPUOpTestWrapper): size_tensor = [] for index, ele in enumerate(self.out_size): size_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs['SizeTensor'] = size_tensor diff --git a/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py index 6e74b437e2d..baeff8a10a6 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_randint_op_xpu.py @@ -35,7 +35,7 @@ def output_hist(out): hist, _ = np.histogram(out, range=(-10, 10)) hist = hist.astype("float32") hist /= float(out.size) - prob = 0.1 * np.ones((10)) + prob = 0.1 * np.ones(10) return hist, prob diff --git a/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py index 4ceaa44c0fb..0e285f6b03c 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_randperm_op_xpu.py @@ -78,7 +78,7 @@ class XPUTestRandpermOp(XPUOpTestWrapper): self.use_xpu = True self.use_mkldnn = False self.inputs = {} - self.outputs = {"Out": np.zeros((self.n)).astype(self.dtype)} + self.outputs = {"Out": np.zeros(self.n).astype(self.dtype)} self.attrs = { "n": self.n, "dtype": convert_dtype(self.dtype), diff --git a/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py index e85ccf0cc44..41415a574b5 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_reshape2_op_xpu.py @@ -118,7 +118,7 @@ class XPUTestReshapeOp(XPUOpTestWrapper): shape_tensor = [] for index, ele in enumerate(self.new_shape): shape_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py index c8d2d6ba2c8..b5284dfb320 100755 --- a/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_rnn_op_xpu.py @@ -97,7 +97,7 @@ class XPUTestRNNOp(XPUOpTestWrapper): self.hidden_size, ) ).astype(self.dtype) - state_out = np.ndarray((300)).astype("uint8") + state_out = np.ndarray(300).astype("uint8") self.inputs = { 'Input': input, @@ -126,7 +126,7 @@ class XPUTestRNNOp(XPUOpTestWrapper): ('last_hidden', last_hidden), ('last_cell', last_cell), ], - 'Reserve': np.ndarray((400)).astype("uint8"), + 'Reserve': np.ndarray(400).astype("uint8"), 'DropoutState': state_out, } diff --git a/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py index bd0d3e9e362..65f52bcfc0b 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_sequence_unpad_op_xpu.py @@ -119,7 +119,7 @@ class TestSequenceUnpadOpError(unittest.TestCase): def test_length_variable(): x1 = paddle.static.data(name='x1', shape=[10, 5], dtype='float32') - len1 = np.random.random((10)).astype("int64") + len1 = np.random.random(10).astype("int64") paddle.static.nn.sequence_lod.sequence_unpad(x=x1, length=len1) self.assertRaises(TypeError, test_length_variable) diff --git a/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py index 7f0982c29ce..c6f9c79be4d 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_tile_op_xpu.py @@ -110,7 +110,7 @@ class XPUTestTileOpRank1_tensor_attr(XPUOpTestWrapper): repeat_times_tensor = [] for index, ele in enumerate(self.repeat_times): repeat_times_tensor.append( - ("x" + str(index), np.ones((1)).astype('int32') * ele) + ("x" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py index 53ab29a322a..56862299074 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_unsqueeze2_op_xpu.py @@ -113,7 +113,7 @@ class XPUTestUnsqueeze2Op(XPUOpTestWrapper): axes_tensor_list = [] for index, ele in enumerate(self.axes): axes_tensor_list.append( - ("axes" + str(index), np.ones((1)).astype('int32') * ele) + ("axes" + str(index), np.ones(1).astype('int32') * ele) ) self.inputs = { diff --git a/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py b/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py index 82c974cd4fe..a7a26f32b02 100644 --- a/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py +++ b/python/paddle/fluid/tests/unittests/xpu/test_where_op_xpu.py @@ -47,7 +47,7 @@ class XPUTestWhereOp(XPUOpTestWrapper): def init_data(self): self.x = np.random.uniform(-3, 5, (100)).astype(self.dtype) self.y = np.random.uniform(-3, 5, (100)).astype(self.dtype) - self.cond = np.zeros((100)).astype("bool") + self.cond = np.zeros(100).astype("bool") def init_config(self): self.op_type = "where" diff --git a/python/paddle/hapi/model.py b/python/paddle/hapi/model.py index e25f88584ac..946766a82a1 100644 --- a/python/paddle/hapi/model.py +++ b/python/paddle/hapi/model.py @@ -1514,7 +1514,7 @@ class Model: except ValueError as err: if skip_mismatch: warnings.warn( - ("Skip loading for {}. ".format(key) + str(err)) + "Skip loading for {}. ".format(key) + str(err) ) # reset optimizer when mismatch happens reset_optimizer = True diff --git a/python/paddle/nn/decode.py b/python/paddle/nn/decode.py index b5e1e287ecf..04d40124016 100644 --- a/python/paddle/nn/decode.py +++ b/python/paddle/nn/decode.py @@ -706,7 +706,7 @@ def _dynamic_decode_imperative( initial_states, initial_finished, ) - cond = paddle.logical_not((paddle.all(initial_finished))) + cond = paddle.logical_not(paddle.all(initial_finished)) sequence_lengths = paddle.cast(paddle.zeros_like(initial_finished), "int64") outputs = None @@ -812,7 +812,7 @@ def _dynamic_decode_declarative( global_finished.stop_gradient = True step_idx = paddle.full(shape=[1], fill_value=0, dtype="int64") - cond = paddle.logical_not((paddle.all(initial_finished))) + cond = paddle.logical_not(paddle.all(initial_finished)) if max_step_num is not None: max_step_num = paddle.full( shape=[1], fill_value=max_step_num, dtype="int64" diff --git a/python/paddle/nn/layer/layers.py b/python/paddle/nn/layer/layers.py index b1f184469fa..23c02148c5b 100644 --- a/python/paddle/nn/layer/layers.py +++ b/python/paddle/nn/layer/layers.py @@ -1896,7 +1896,7 @@ class Layer: match_res = _check_match(key_name, param) matched_param_state.append(match_res) except ValueError as err: - warnings.warn(("Skip loading for {}. ".format(key) + str(err))) + warnings.warn("Skip loading for {}. ".format(key) + str(err)) for key in state_dict.keys(): if key not in match_keys: unexpected_keys.append(key) diff --git a/test/legacy_test/test_transforms_static.py b/test/legacy_test/test_transforms_static.py index 8e6c14635c2..98bc132486c 100644 --- a/test/legacy_test/test_transforms_static.py +++ b/test/legacy_test/test_transforms_static.py @@ -126,7 +126,7 @@ class TestRandomCrop_random(TestTransformUnitTestBase): for y in range(h - c_h): for x in range(w - c_w): diff_abs_sum = np.abs( - (self.img[:, y : y + c_h, x : x + c_w] - res) + self.img[:, y : y + c_h, x : x + c_w] - res ).sum() if diff_abs_sum < eps: res_assert = False diff --git a/tools/get_pr_ut.py b/tools/get_pr_ut.py index 3377bf9f7fc..91b3cb4d927 100644 --- a/tools/get_pr_ut.py +++ b/tools/get_pr_ut.py @@ -162,7 +162,7 @@ class PRChecker: ) if 'cmakelist' in filename.lower(): isWhiteFile = False - elif filename.startswith((not_white_files)): + elif filename.startswith(not_white_files): isWhiteFile = False else: isWhiteFile = True @@ -330,7 +330,7 @@ class PRChecker: if filename.startswith(PADDLE_ROOT + 'python/'): file_list.append(filename) elif filename.startswith(PADDLE_ROOT + 'paddle/'): - if filename.startswith((PADDLE_ROOT + 'paddle/infrt')): + if filename.startswith(PADDLE_ROOT + 'paddle/infrt'): filterFiles.append(filename) elif filename.startswith(PADDLE_ROOT + 'paddle/scripts'): if filename.startswith( diff --git a/tools/get_single_test_cov.py b/tools/get_single_test_cov.py index 415d43e0b01..ba95a59fc1a 100644 --- a/tools/get_single_test_cov.py +++ b/tools/get_single_test_cov.py @@ -48,8 +48,8 @@ def getFNDAFile(rootPath, test): os.system('echo %s >> %s' % (message_list[1], fn_filename)) if 'FNH:0' not in gcov_data: for message in message_list: - if message.startswith(('FNDA:')) and ( - not message.startswith(('FNDA:0,')) + if message.startswith('FNDA:') and ( + not message.startswith('FNDA:0,') ): tmp_data = message.split('FNDA:')[1].split(',') hit = int(tmp_data[0]) @@ -159,8 +159,8 @@ def getBaseFnda(rootPath, test): # only for cc file if ".cc" in message_list[1]: for message in message_list: - if message.startswith(('FNDA:')) and ( - not message.startswith(('FNDA:0,')) + if message.startswith('FNDA:') and ( + not message.startswith('FNDA:0,') ): tmp_data = message.split('FNDA:')[1].split(',') symbol_fnda[tmp_data[1]] = int(tmp_data[0]) -- GitLab