From 8fe94d449e65cc25aa3b685ca5de0293a694d3ae Mon Sep 17 00:00:00 2001 From: bukejiyu <52310069+bukejiyu@users.noreply.github.com> Date: Mon, 17 Jul 2023 17:16:24 +0800 Subject: [PATCH] [Inference][trt] fix trt ci and timeout (#55441) * unary bitwise_not adapter tensorRT8.6 in Paddle-TensorRT * Update test_trt_inspector.py * test_trt_convert_conv2d_transpose * Update test_trt_convert_conv2d_transpose.py --- test/ir/inference/test_trt_convert_conv2d_transpose.py | 9 +++++++-- .../inference/test_trt_convert_cross_multihead_matmul.py | 8 ++++---- test/ir/inference/test_trt_int64.py | 5 ++++- 3 files changed, 15 insertions(+), 7 deletions(-) diff --git a/test/ir/inference/test_trt_convert_conv2d_transpose.py b/test/ir/inference/test_trt_convert_conv2d_transpose.py index 98e6597540d..42057072f70 100644 --- a/test/ir/inference/test_trt_convert_conv2d_transpose.py +++ b/test/ir/inference/test_trt_convert_conv2d_transpose.py @@ -252,10 +252,15 @@ class TrtConvertConv2dTransposeTest2(TrtLayerAutoScanTest): self.trt_param.workspace_size = 1073741824 def generate_input1(batch, num_channels, attrs: List[Dict[str, Any]]): - return np.ones([batch, num_channels, 20, 30]).astype(np.float32) + return ( + np.ones([batch, num_channels, 20, 30]).astype(np.float32) / 100 + ) def generate_weight1(num_channels, attrs: List[Dict[str, Any]]): - return np.random.random([num_channels, 64, 3, 3]).astype(np.float32) + return ( + np.random.random([num_channels, 64, 3, 3]).astype(np.float32) + / 100 + ) num_channels = 128 batch = 1 diff --git a/test/ir/inference/test_trt_convert_cross_multihead_matmul.py b/test/ir/inference/test_trt_convert_cross_multihead_matmul.py index c427a94772b..a8ce9ee0a59 100644 --- a/test/ir/inference/test_trt_convert_cross_multihead_matmul.py +++ b/test/ir/inference/test_trt_convert_cross_multihead_matmul.py @@ -247,15 +247,15 @@ class TrtConvertCrossMultiHeadMatmulTest(TrtLayerAutoScanTest): # The last dim of input1 and input2 should be static. self.dynamic_shape.min_input_shape = { "input_data1": [1, 4096, 320], - "input_data2": [1, 77, 768], + "input_data2": [1, 768, 768], } self.dynamic_shape.max_input_shape = { "input_data1": [8, 4096, 320], - "input_data2": [8, 77, 768], + "input_data2": [8, 768, 768], } self.dynamic_shape.opt_input_shape = { "input_data1": [2, 4096, 320], - "input_data2": [2, 77, 768], + "input_data2": [2, 768, 768], } def clear_dynamic_shape(): @@ -281,7 +281,7 @@ class TrtConvertCrossMultiHeadMatmulTest(TrtLayerAutoScanTest): self.trt_param.workspace_size = 2013265920 yield self.create_inference_config(), (1, 3), (1e-5, 1e-4) self.trt_param.precision = paddle_infer.PrecisionType.Half - yield self.create_inference_config(), (1, 3), (1e-2, 1e-3) + yield self.create_inference_config(), (1, 3), (1e-2, 1e-2) def add_skip_trt_case(self): def teller1(program_config, predictor_config): diff --git a/test/ir/inference/test_trt_int64.py b/test/ir/inference/test_trt_int64.py index f83cffa3400..2ea5aef705b 100644 --- a/test/ir/inference/test_trt_int64.py +++ b/test/ir/inference/test_trt_int64.py @@ -78,7 +78,6 @@ class TrtInt64Test1(TrtLayerAutoScanTest): "infer_flags": infer_flags, } ] - ops_config = [ { "op_type": "slice", @@ -87,6 +86,9 @@ class TrtInt64Test1(TrtLayerAutoScanTest): "Out": ["slice_output_data"] }, "op_attrs": dics[0], + "outputs_dtype": { + "slice_output_data": np.int64 + }, } ] ops = self.generate_op_config(ops_config) @@ -162,6 +164,7 @@ class TrtInt64Test2(TrtLayerAutoScanTest): }, "op_outputs": {"Out": ["output_data"]}, "op_attrs": dics[0], + "outputs_dtype": {"slice_output_data": np.int64}, } ] ops = self.generate_op_config(ops_config) -- GitLab