diff --git a/paddle/fluid/inference/api/api_tensorrt_subgraph_engine.cc b/paddle/fluid/inference/api/api_tensorrt_subgraph_engine.cc index 8be444949655ed404179ca786245cbd95f55db4e..ce147eb5da77824930ed1a2eb609766b24ced598 100644 --- a/paddle/fluid/inference/api/api_tensorrt_subgraph_engine.cc +++ b/paddle/fluid/inference/api/api_tensorrt_subgraph_engine.cc @@ -90,7 +90,7 @@ class TensorRTSubgraphPredictor : public NativePaddlePredictor { // Analyze inference_program Argument argument; - int* minimum_subgraph_size = new int(config_.minimun_subgraph_size); + int* minimum_subgraph_size = new int(config_.minimum_subgraph_size); int* max_batch_size = new int(config_.max_batch_size); int* workspace_size = new int(config_.workspace_size); std::string* precision_mode = new std::string(config_.precision_mode); diff --git a/paddle/fluid/inference/api/paddle_inference_api.h b/paddle/fluid/inference/api/paddle_inference_api.h index 084da823e07e0130a1ba6f702b4a5ab849fdd43b..bacb319e24b067cdd38618128825c34f11df3b2d 100644 --- a/paddle/fluid/inference/api/paddle_inference_api.h +++ b/paddle/fluid/inference/api/paddle_inference_api.h @@ -155,7 +155,7 @@ struct TensorRTConfig : public NativeConfig { // We set this variable to control the minimum number of nodes in the // subgraph, 3 as // default value. - int minimun_subgraph_size = 3; + int minimum_subgraph_size = 3; // Reserved configuration // We just support "FP32" now, "FP16" and "INT8" will be supported. std::string precision_mode = "FP32";