diff --git a/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_cn.rst b/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_cn.rst index c1bfba460db6c12651ac6a04f823812642490c9f..9e255c3d92992c521262b9f78f157b247ab25631 100644 --- a/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_cn.rst +++ b/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_cn.rst @@ -7,16 +7,16 @@ ------------- .. csv-table:: - :header: "版本说明", "预测库(1.8.3版本)", "预测库(develop版本)" - :widths: 3, 2, 2 + :header: "版本说明", "预测库(1.8.4版本)", "预测库(2.0.0-beta0版本)", "预测库(develop版本)" + :widths: 3, 2, 2, 2 - "ubuntu14.04_cpu_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cpu_avx_openblas", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cpu_noavx_openblas", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda9.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda10.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda10.1_cudnn7.6_avx_mkl_trt6", "`fluid_inference.tgz `_", - "nv-jetson-cuda10-cudnn7.5-trt5", "`fluid_inference.tar.gz `_", + "ubuntu14.04_cpu_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cpu_avx_openblas", "`fluid_inference.tgz `_", ,"`paddle_inference.tgz `_" + "ubuntu14.04_cpu_noavx_openblas", "`fluid_inference.tgz `_", ,"`paddle_inference.tgz `_" + "ubuntu14.04_cuda9.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cuda10.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cuda10.1_cudnn7.6_avx_mkl_trt6", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", + "nv-jetson-cuda10-cudnn7.5-trt5", "`fluid_inference.tar.gz `_", "`paddle_inference.tar.gz `_", 从源码编译 @@ -46,7 +46,7 @@ WITH_NV_JETSON OFF 在NV Jetson硬件上编译时需 git clone https://github.com/paddlepaddle/Paddle cd Paddle # 建议使用git checkout切换到Paddle稳定的版本,如: - git checkout v1.7.2 + git checkout v1.8.4 **note**: 如果您是多卡机器,建议安装NCCL;如果您是单卡机器则可以在编译时显示指定WITH_NCCL=OFF来跳过这一步。注意如果WITH_NCCL=ON,且没有安装NCCL,则编译会报错。 @@ -188,3 +188,4 @@ version.txt 中记录了该预测库的版本信息,包括Git Commit ID、使 WITH_GPU: ON CUDA version: 10.1 CUDNN version: v7 + diff --git a/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_en.rst b/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_en.rst index 545aba61360b0018e3d3a1c28f4e56f4f6005925..60d19b0906f002191fd76b00acfd023c0618420d 100644 --- a/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_en.rst +++ b/doc/paddle/advanced_guide/inference_deployment/inference/build_and_install_lib_en.rst @@ -7,16 +7,16 @@ Direct Download and Installation --------------------------------- .. csv-table:: c++ inference library list - :header: "version description", "inference library(1.8.3 version)", "inference library(develop version)" - :widths: 3, 2, 2 + :header: "version description", "inference library(1.8.4 version)", "inference library(2.0.0-beta0 version)", "inference library(develop version)" + :widths: 3, 2, 2, 2 - "ubuntu14.04_cpu_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cpu_avx_openblas", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cpu_noavx_openblas", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda9.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda10.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`fluid_inference.tgz `_" - "ubuntu14.04_cuda10.1_cudnn7.6_avx_mkl_trt6", "`fluid_inference.tgz `_", - "nv-jetson-cuda10-cudnn7.5-trt5", "`fluid_inference.tar.gz `_", + "ubuntu14.04_cpu_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cpu_avx_openblas", "`fluid_inference.tgz `_", ,"`paddle_inference.tgz `_" + "ubuntu14.04_cpu_noavx_openblas", "`fluid_inference.tgz `_", ,"`paddle_inference.tgz `_" + "ubuntu14.04_cuda9.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cuda10.0_cudnn7_avx_mkl", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", "`paddle_inference.tgz `_" + "ubuntu14.04_cuda10.1_cudnn7.6_avx_mkl_trt6", "`fluid_inference.tgz `_", "`paddle_inference.tgz `_", + "nv-jetson-cuda10-cudnn7.5-trt5", "`fluid_inference.tar.gz `_", "`paddle_inference.tar.gz `_", Build from Source Code ----------------------- @@ -46,8 +46,8 @@ Firstly we pull the latest code from github. git clone https://github.com/paddlepaddle/Paddle cd Paddle - # Use git checkout to switch to stable versions such as v1.7.2 - git checkout v1.7.2 + # Use git checkout to switch to stable versions such as v1.8.4 + git checkout v1.8.4 **note**: If your environment is a multi-card machine, it is recommended to install nccl; otherwise, you can skip this step by specifying WITH_NCCL = OFF during compilation. Note that if WITH_NCCL = ON, and NCCL is not installed, the compiler will report an error. @@ -202,3 +202,4 @@ The version information of the inference library is recorded in version.txt, inc WITH_GPU: ON CUDA version: 8.0 CUDNN version: v7 +