diff --git a/paddle/fluid/inference/api/analysis_config.cc b/paddle/fluid/inference/api/analysis_config.cc index 853c1ac1da8742733e609c1dea098a208eadc015..b5ca0ef5924397544882741078043d747a145ebf 100644 --- a/paddle/fluid/inference/api/analysis_config.cc +++ b/paddle/fluid/inference/api/analysis_config.cc @@ -421,7 +421,6 @@ void AnalysisConfig::Update() { pass_builder()->AppendPass(pass); } } - LOG(INFO) << "use_dlnne_:" << use_dlnne_ << std::endl; if (use_dlnne_) { pass_builder()->ClearPasses(); for (const auto &pass : kDlnneSubgraphPasses) { diff --git a/paddle/fluid/inference/api/analysis_predictor.cc b/paddle/fluid/inference/api/analysis_predictor.cc index 215174c12ce3b2d5c5f432c2f4ad7a0ca48c76b6..b205d553c99ca5cdadef664bb20e23ebba431d45 100644 --- a/paddle/fluid/inference/api/analysis_predictor.cc +++ b/paddle/fluid/inference/api/analysis_predictor.cc @@ -152,8 +152,8 @@ bool AnalysisPredictor::Init( : platform::ProfilerState::kCPU; platform::EnableProfiler(tracking_device); } else { - LOG(INFO) << "Profiler is deactivated, and no profiling report will be " - "generated."; + VLOG(2) << "Profiler is deactivated, and no profiling report will be " + "generated."; } // no matter with or without MKLDNN diff --git a/paddle/fluid/inference/api/paddle_analysis_config.h b/paddle/fluid/inference/api/paddle_analysis_config.h index 2bbd4bb837a22f672e5aa625f299424b6f0c5b88..f9e4869934a0fe84647e7a81a9e36f9fd03d7c42 100644 --- a/paddle/fluid/inference/api/paddle_analysis_config.h +++ b/paddle/fluid/inference/api/paddle_analysis_config.h @@ -294,7 +294,7 @@ struct PD_INFER_DECL AnalysisConfig { /// workspace. /// \param max_batch_size The maximum batch size of this prediction task, /// better set as small as possible for less performance loss. - /// \param min_subgrpah_size The minimum TensorRT subgraph size needed, if a + /// \param min_subgraph_size The minimum TensorRT subgraph size needed, if a /// subgraph is smaller than this, it will not be transferred to TensorRT /// engine. /// \param precision The precision used in TensorRT. diff --git a/paddle/fluid/inference/tensorrt/plugin/emb_eltwise_layernorm_plugin.cu b/paddle/fluid/inference/tensorrt/plugin/emb_eltwise_layernorm_plugin.cu index 6d3872aaeb8a77acf1455e4d5e555ee01d36478a..c873b1fc310def0e4cab2061a170a70055f1d9f2 100644 --- a/paddle/fluid/inference/tensorrt/plugin/emb_eltwise_layernorm_plugin.cu +++ b/paddle/fluid/inference/tensorrt/plugin/emb_eltwise_layernorm_plugin.cu @@ -134,7 +134,7 @@ int EmbEltwiseLayernormPluginDynamicImpl::enqueue( int batch = id_dims.d[0]; int seq_len = id_dims.d[1]; int input_num = embs_.size(); - + cudaGetDevice(&device_id_); auto in_ptr_gpu_d = in_ptr_tensor_.mutable_data(platform::CUDAPlace(device_id_)); auto emb_ptr_gpu_d = diff --git a/paddle/fluid/inference/tests/api/trt_dynamic_shape_ernie_test.cc b/paddle/fluid/inference/tests/api/trt_dynamic_shape_ernie_test.cc index 45dff9f4c371020df651aee31d23010acb61cad7..a45b78f05e73c48e5fb378b77753555a9fdb64b8 100644 --- a/paddle/fluid/inference/tests/api/trt_dynamic_shape_ernie_test.cc +++ b/paddle/fluid/inference/tests/api/trt_dynamic_shape_ernie_test.cc @@ -29,11 +29,6 @@ void run(const AnalysisConfig& config, std::vector* out_data) { int run_batch = 1; const int run_seq_len = 128; - std::vector tmp_input; - std::vector tmp_four_input; - tmp_input.reserve(run_batch * run_seq_len); - tmp_four_input.reserve(run_batch * run_seq_len); - int64_t i0[run_seq_len] = { 1, 3558, 4, 75, 491, 89, 340, 313, 93, 4, 255, 10, 75, 321, 4095, 1902, 4, 134, 49, 75, 311, 14, 44, 178, 543, 15, 12043, 2,