diff --git a/doc/fluid/advanced_guide/inference_deployment/inference/native_infer.md b/doc/fluid/advanced_guide/inference_deployment/inference/native_infer.md index a360b0cd774f311c4dbaabd8c6575f033a6b2e91..e2fac6526eec1895d690b81e14301d752e2a0216 100644 --- a/doc/fluid/advanced_guide/inference_deployment/inference/native_infer.md +++ b/doc/fluid/advanced_guide/inference_deployment/inference/native_infer.md @@ -228,7 +228,7 @@ float *output_d = output_t->data(PaddlePlace::kGPU, &output_size); 3. 在CPU可用核心数足够时,可以将设置`config->SetCpuMathLibraryNumThreads(num);`中的num值调高一些。 ### GPU下预测 -1. 可以尝试打开 TensorRT 子图加速引擎, 通过计算图分析,Paddle可以自动将计算图中部分子图融合,并调用NVIDIA的 TensorRT 来进行加速,详细内容可以参考 [使用Paddle-TensorRT库预测](./paddle_tensorrt_infer.html)。 +1. 可以尝试打开 TensorRT 子图加速引擎, 通过计算图分析,Paddle可以自动将计算图中部分子图融合,并调用NVIDIA的 TensorRT 来进行加速,详细内容可以参考 [使用Paddle-TensorRT库预测](../../performance_improving/inference_improving/paddle_tensorrt_infer.html)。 ### 多线程预测 Paddle Fluid支持通过在不同线程运行多个AnalysisPredictor的方式来优化预测性能,支持CPU和GPU环境。 diff --git a/doc/fluid/advanced_guide/inference_deployment/inference/native_infer_en.md b/doc/fluid/advanced_guide/inference_deployment/inference/native_infer_en.md index 4a8b012dee0b460248f8f9ebf5dba8b518ca6ce6..69cc7bf0818f7dc5989a561b29fee6a10c847c6c 100644 --- a/doc/fluid/advanced_guide/inference_deployment/inference/native_infer_en.md +++ b/doc/fluid/advanced_guide/inference_deployment/inference/native_infer_en.md @@ -214,7 +214,7 @@ float *output_d = output_t->data(PaddlePlace::kGPU, &output_size); 3. When the number of CPU cores available is enough, you can increase the num value in the setting `config->SetCpuMathLibraryNumThreads(num);`. ### Tuning on GPU -1. You can try to open the TensorRT subgraph acceleration engine. Through the graph analysis, Paddle can automatically fuse certain subgraphs, and call NVIDIA's TensorRT for acceleration. For details, please refer to [Use Paddle-TensorRT Library for inference](./paddle_tensorrt_infer_en.html)。 +1. You can try to open the TensorRT subgraph acceleration engine. Through the graph analysis, Paddle can automatically fuse certain subgraphs, and call NVIDIA's TensorRT for acceleration. For details, please refer to [Use Paddle-TensorRT Library for inference](../../performance_improving/inference_improving/paddle_tensorrt_infer_en.html)。 ### Tuning with multi-thread Paddle Fluid supports optimizing prediction performance by running multiple AnalysisPredictors on different threads, and supports CPU and GPU environments. diff --git a/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer.md b/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer.md index 6ccdedbeae58859528a79f2659575146047b71e3..f6a7ff48487430c4ea37b99550c8cc1be5321f62 100644 --- a/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer.md +++ b/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer.md @@ -59,7 +59,7 @@ config->EnableTensorRtEngine(1 << 20 /* workspace_size*/, ## Paddle-TRT样例编译测试 -1. 下载或编译带有 TensorRT 的paddle预测库,参考[安装与编译C++预测库](./build_and_install_lib_cn.html)。 +1. 下载或编译带有 TensorRT 的paddle预测库,参考[安装与编译C++预测库](../../inference_deployment/inference/build_and_install_lib_cn.html)。 2. 从[NVIDIA官网](https://developer.nvidia.com/nvidia-tensorrt-download)下载对应本地环境中cuda和cudnn版本的TensorRT,需要登陆NVIDIA开发者账号。 3. 下载[预测样例](https://paddle-inference-dist.bj.bcebos.com/tensorrt_test/paddle_inference_sample_v1.7.tar.gz)并解压,进入`sample/paddle-TRT`目录下。 diff --git a/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer_en.md b/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer_en.md index 83f41bb81c5a2a1caff6b527cbb4794d400b9b01..f3e7d7961b705ebb0be634ff03ead6e1f464db95 100644 --- a/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer_en.md +++ b/doc/fluid/advanced_guide/performance_improving/inference_improving/paddle_tensorrt_infer_en.md @@ -52,7 +52,7 @@ The details of this interface is as following: ## Paddle-TRT example compiling test -1. Download or compile Paddle Inference with TensorRT support, refer to [Install and Compile C++ Inference Library](./build_and_install_lib_en.html). +1. Download or compile Paddle Inference with TensorRT support, refer to [Install and Compile C++ Inference Library](../../inference_deployment/inference/build_and_install_lib_en.html). 2. Download NVIDIA TensorRT(with consistent version of cuda and cudnn in local environment) from [NVIDIA TensorRT](https://developer.nvidia.com/nvidia-tensorrt-download) with an NVIDIA developer account. 3. Download [Paddle Inference sample](https://paddle-inference-dist.bj.bcebos.com/tensorrt_test/paddle_inference_sample_v1.7.tar.gz) and uncompress, and enter `sample/paddle-TRT` directory.