提交 51ae50ce 编写于 作者: S sheqiZ 提交者: xsrobin

TRT预测库仅支持GPU编译 (#901)

* TRT预测库仅支持GPU编译

DLTP-739 完善tensorRT使用文档 TRT需要GPU编译

* DLTP-965 预测模型的保存和加载文档优化

* Windows_installation_English

* DLTP-1417 受训改为训练
上级 921765ab
...@@ -12,20 +12,22 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习 ...@@ -12,20 +12,22 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习
## <a name="编译Paddle-TRT预测库">编译Paddle-TRT预测库</a> ## <a name="编译Paddle-TRT预测库">编译Paddle-TRT预测库</a>
**使用Docker编译预测库** **使用Docker编译预测库**
TensorRT预测库目前仅支持使用GPU编译。
1. 下载Paddle 1. 下载Paddle
``` ```
git clone https://github.com/PaddlePaddle/Paddle.git git clone https://github.com/PaddlePaddle/Paddle.git
``` ```
2. 获取docker镜像 2. 获取docker镜像
``` ```
nvidia-docker run --name paddle_trt -v $PWD/Paddle:/Paddle -it hub.baidubce.com/paddlepaddle/paddle:latest-dev /bin/bash nvidia-docker run --name paddle_trt -v $PWD/Paddle:/Paddle -it hub.baidubce.com/paddlepaddle/paddle:latest-dev /bin/bash
``` ```
3. 编译Paddle TensorRT 3. 编译Paddle TensorRT
``` ```
...@@ -43,15 +45,15 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习 ...@@ -43,15 +45,15 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习
-DWITH_PYTHON=OFF \ -DWITH_PYTHON=OFF \
-DTENSORRT_ROOT=/usr \ -DTENSORRT_ROOT=/usr \
-DON_INFER=ON -DON_INFER=ON
# 编译 # 编译
make -j make -j
# 生成预测库 # 生成预测库
make inference_lib_dist -j make inference_lib_dist -j
``` ```
编译后的库的目录如下: 编译后的库的目录如下:
``` ```
fluid_inference_install_dir fluid_inference_install_dir
├── paddle ├── paddle
...@@ -61,12 +63,12 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习 ...@@ -61,12 +63,12 @@ NVIDIA TensorRT 是一个高性能的深度学习预测库,可为深度学习
├── third_party ├── third_party
├── boost ├── boost
├── install ├── install
└── engine3 └── engine3
``` ```
`fluid_inference_install_dir`下, paddle目录包含了预测库的头文件和预测库的lib, version.txt 中包含了lib的版本和配置信息,third_party 中包含了预测库依赖的第三方库 `fluid_inference_install_dir`下, paddle目录包含了预测库的头文件和预测库的lib, version.txt 中包含了lib的版本和配置信息,third_party 中包含了预测库依赖的第三方库
## <a name="Paddle-TRT接口使用">Paddle-TRT接口使用</a> ## <a name="Paddle-TRT接口使用">Paddle-TRT接口使用</a>
[`paddle_inference_api.h`]('https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/inference/api/paddle_inference_api.h') 定义了使用TensorRT的所有接口。 [`paddle_inference_api.h`]('https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/inference/api/paddle_inference_api.h') 定义了使用TensorRT的所有接口。
...@@ -89,13 +91,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) { ...@@ -89,13 +91,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) {
AnalysisConfig config(model_dirname); AnalysisConfig config(model_dirname);
// config->SetModel(model_dirname + "/model", // config->SetModel(model_dirname + "/model",
// model_dirname + "/params"); // model_dirname + "/params");
config->EnableUseGpu(100, 0 /*gpu_id*/); config->EnableUseGpu(100, 0 /*gpu_id*/);
config->EnableTensorRtEngine(1 << 20 /*work_space_size*/, batch_size /*max_batch_size*/); config->EnableTensorRtEngine(1 << 20 /*work_space_size*/, batch_size /*max_batch_size*/);
// 2. 根据config 创建predictor // 2. 根据config 创建predictor
auto predictor = CreatePaddlePredictor(config); auto predictor = CreatePaddlePredictor(config);
// 3. 创建输入 tensor // 3. 创建输入 tensor
int height = 224; int height = 224;
int width = 224; int width = 224;
float data[batch_size * 3 * height * width] = {0}; float data[batch_size * 3 * height * width] = {0};
...@@ -114,13 +116,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) { ...@@ -114,13 +116,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) {
const size_t num_elements = outputs.front().data.length() / sizeof(float); const size_t num_elements = outputs.front().data.length() / sizeof(float);
auto *data = static_cast<float *>(outputs.front().data.data()); auto *data = static_cast<float *>(outputs.front().data.data());
for (size_t i = 0; i < num_elements; i++) { for (size_t i = 0; i < num_elements; i++) {
std::cout << "output: " << data[i] << std::endl; std::cout << "output: " << data[i] << std::endl;
} }
} }
} // namespace paddle } // namespace paddle
int main() { int main() {
// 模型下载地址 http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/mobilenet.tar.gz // 模型下载地址 http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/mobilenet.tar.gz
paddle::RunTensorRT(1, "./mobilenet"); paddle::RunTensorRT(1, "./mobilenet");
return 0; return 0;
...@@ -133,9 +135,9 @@ int main() { ...@@ -133,9 +135,9 @@ int main() {
``` ```
wget http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/paddle_trt_samples.tar.gz wget http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/paddle_trt_samples.tar.gz
``` ```
解压后的目录如下: 解压后的目录如下:
``` ```
sample sample
├── CMakeLists.txt ├── CMakeLists.txt
...@@ -146,12 +148,12 @@ int main() { ...@@ -146,12 +148,12 @@ int main() {
│ └── params │ └── params
└── run_impl.sh └── run_impl.sh
``` ```
- `mobilenet_test.cc` 为单线程的程序文件 - `mobilenet_test.cc` 为单线程的程序文件
- `thread_mobilenet_test.cc` 为多线程的程序文件 - `thread_mobilenet_test.cc` 为多线程的程序文件
- `mobilenetv1` 为模型文件 - `mobilenetv1` 为模型文件
在这里假设预测库的路径为 ``BASE_DIR/fluid_inference_install_dir/`` ,样例所在的目录为 ``SAMPLE_BASE_DIR/sample`` 在这里假设预测库的路径为 ``BASE_DIR/fluid_inference_install_dir/`` ,样例所在的目录为 ``SAMPLE_BASE_DIR/sample``
2. 编译样例 2. 编译样例
...@@ -181,10 +183,10 @@ int main() { ...@@ -181,10 +183,10 @@ int main() {
# sh run_impl.sh {预测库的地址} {测试脚本的名字} {模型目录} # sh run_impl.sh {预测库的地址} {测试脚本的名字} {模型目录}
# 我们随机生成了500个输入来模拟这一过程,建议大家用真实样例进行实验。 # 我们随机生成了500个输入来模拟这一过程,建议大家用真实样例进行实验。
sh run_impl.sh BASE_DIR/fluid_inference_install_dir/ fluid_generate_calib_test SAMPLE_BASE_DIR/sample/mobilenetv1 sh run_impl.sh BASE_DIR/fluid_inference_install_dir/ fluid_generate_calib_test SAMPLE_BASE_DIR/sample/mobilenetv1
``` ```
运行结束后,在 `SAMPLE_BASE_DIR/sample/build/mobilenetv1` 模型目录下会多出一个名字为trt_calib_*的文件,即校准表。 运行结束后,在 `SAMPLE_BASE_DIR/sample/build/mobilenetv1` 模型目录下会多出一个名字为trt_calib_*的文件,即校准表。
``` shell ``` shell
# 执行INT8预测 # 执行INT8预测
# 将带校准表的模型文件拷贝到特定地址 # 将带校准表的模型文件拷贝到特定地址
...@@ -193,7 +195,7 @@ int main() { ...@@ -193,7 +195,7 @@ int main() {
``` ```
## <a name="Paddle-TRT子图运行原理">Paddle-TRT子图运行原理</a> ## <a name="Paddle-TRT子图运行原理">Paddle-TRT子图运行原理</a>
PaddlePaddle采用子图的形式对TensorRT进行集成,当模型加载后,神经网络可以表示为由变量和运算节点组成的计算图。Paddle TensorRT实现的功能是能够对整个图进行扫描,发现图中可以使用TensorRT优化的子图,并使用TensorRT节点替换它们。在模型的推断期间,如果遇到TensorRT节点,Paddle会调用TensoRT库对该节点进行优化,其他的节点调用Paddle的原生实现。TensorRT在推断期间能够进行Op的横向和纵向融合,过滤掉冗余的Op,并对特定平台下的特定的Op选择合适的kenel等进行优化,能够加快模型的预测速度。 PaddlePaddle采用子图的形式对TensorRT进行集成,当模型加载后,神经网络可以表示为由变量和运算节点组成的计算图。Paddle TensorRT实现的功能是能够对整个图进行扫描,发现图中可以使用TensorRT优化的子图,并使用TensorRT节点替换它们。在模型的推断期间,如果遇到TensorRT节点,Paddle会调用TensoRT库对该节点进行优化,其他的节点调用Paddle的原生实现。TensorRT在推断期间能够进行Op的横向和纵向融合,过滤掉冗余的Op,并对特定平台下的特定的Op选择合适的kenel等进行优化,能够加快模型的预测速度。
下图使用一个简单的模型展示了这个过程: 下图使用一个简单的模型展示了这个过程:
...@@ -208,12 +210,5 @@ int main() { ...@@ -208,12 +210,5 @@ int main() {
<img src="https://raw.githubusercontent.com/NHZlX/FluidDoc/add_trt_doc/doc/fluid/user_guides/howto/inference/image/model_graph_trt.png" width="600"> <img src="https://raw.githubusercontent.com/NHZlX/FluidDoc/add_trt_doc/doc/fluid/user_guides/howto/inference/image/model_graph_trt.png" width="600">
</p> </p>
我们可以在原始模型网络中看到,绿色节点表示可以被TensorRT支持的节点,红色节点表示网络中的变量,黄色表示Paddle只能被Paddle原生实现执行的节点。那些在原始网络中的绿色节点被提取出来汇集成子图,并由一个TensorRT节点代替,成为转换网络中的`block-25` 节点。在网络运行过程中,如果遇到该节点,Paddle将调用TensorRT库来对其执行。
我们可以在原始模型网络中看到,绿色节点表示可以被TensorRT支持的节点,红色节点表示网络中的变量,黄色表示Paddle只能被Paddle原生实现执行的节点。那些在原始网络中的绿色节点被提取出来汇集成子图,并由一个TensorRT节点代替,成为转换网络中的`block-25` 节点。在网络运行过程中,如果遇到该节点,Paddle将调用TensorRT库来对其执行。
...@@ -9,23 +9,25 @@ Subgraph is used in PaddlePaddle to preliminarily integrate TensorRT, which enab ...@@ -9,23 +9,25 @@ Subgraph is used in PaddlePaddle to preliminarily integrate TensorRT, which enab
- [Paddle-TRT example compiling test](#Paddle-TRT example compiling test) - [Paddle-TRT example compiling test](#Paddle-TRT example compiling test)
- [Paddle-TRT INT8 usage](#Paddle-TRT_INT8 usage) - [Paddle-TRT INT8 usage](#Paddle-TRT_INT8 usage)
- [Paddle-TRT subgraph operation principle](#Paddle-TRT subgraph operation principle) - [Paddle-TRT subgraph operation principle](#Paddle-TRT subgraph operation principle)
## <a name="compile Paddle-TRT inference libraries">compile Paddle-TRT inference libraries</a> ## <a name="compile Paddle-TRT inference libraries">compile Paddle-TRT inference libraries</a>
**Use Docker to build inference libraries** **Use Docker to build inference libraries**
TRT inference libraries can only be compiled using GPU.
1. Download Paddle 1. Download Paddle
``` ```
git clone https://github.com/PaddlePaddle/Paddle.git git clone https://github.com/PaddlePaddle/Paddle.git
``` ```
2. Get docker image 2. Get docker image
``` ```
nvidia-docker run --name paddle_trt -v $PWD/Paddle:/Paddle -it hub.baidubce.com/paddlepaddle/paddle:latest-dev /bin/bash nvidia-docker run --name paddle_trt -v $PWD/Paddle:/Paddle -it hub.baidubce.com/paddlepaddle/paddle:latest-dev /bin/bash
``` ```
3. Build Paddle TensorRT 3. Build Paddle TensorRT
``` ```
...@@ -41,16 +43,16 @@ Subgraph is used in PaddlePaddle to preliminarily integrate TensorRT, which enab ...@@ -41,16 +43,16 @@ Subgraph is used in PaddlePaddle to preliminarily integrate TensorRT, which enab
-DWITH_PYTHON=OFF \ -DWITH_PYTHON=OFF \
-DTENSORRT_ROOT=/usr \ -DTENSORRT_ROOT=/usr \
-DON_INFER=ON -DON_INFER=ON
# build # build
make -j make -j
# generate inference library # generate inference library
make inference_lib_dist -j make inference_lib_dist -j
``` ```
## <a name="Paddle-TRT interface usage">Paddle-TRT interface usage</a> ## <a name="Paddle-TRT interface usage">Paddle-TRT interface usage</a>
[`paddle_inference_api.h`]('https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/inference/api/paddle_inference_api.h') defines all APIs of TensorRT. [`paddle_inference_api.h`]('https://github.com/PaddlePaddle/Paddle/blob/develop/paddle/fluid/inference/api/paddle_inference_api.h') defines all APIs of TensorRT.
General steps are as follows: General steps are as follows:
1. Create appropriate AnalysisConfig. 1. Create appropriate AnalysisConfig.
...@@ -71,13 +73,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) { ...@@ -71,13 +73,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) {
AnalysisConfig config(model_dirname); AnalysisConfig config(model_dirname);
// config->SetModel(model_dirname + "/model", // config->SetModel(model_dirname + "/model",
// model_dirname + "/params"); // model_dirname + "/params");
config->EnableUseGpu(100, 0 /*gpu_id*/); config->EnableUseGpu(100, 0 /*gpu_id*/);
config->EnableTensorRtEngine(1 << 20 /*work_space_size*/, batch_size /*max_batch_size*/); config->EnableTensorRtEngine(1 << 20 /*work_space_size*/, batch_size /*max_batch_size*/);
// 2. Create predictor based on config // 2. Create predictor based on config
auto predictor = CreatePaddlePredictor(config); auto predictor = CreatePaddlePredictor(config);
// 3. Create input tensor // 3. Create input tensor
int height = 224; int height = 224;
int width = 224; int width = 224;
float data[batch_size * 3 * height * width] = {0}; float data[batch_size * 3 * height * width] = {0};
...@@ -96,13 +98,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) { ...@@ -96,13 +98,13 @@ void RunTensorRT(int batch_size, std::string model_dirname) {
const size_t num_elements = outputs.front().data.length() / sizeof(float); const size_t num_elements = outputs.front().data.length() / sizeof(float);
auto *data = static_cast<float *>(outputs.front().data.data()); auto *data = static_cast<float *>(outputs.front().data.data());
for (size_t i = 0; i < num_elements; i++) { for (size_t i = 0; i < num_elements; i++) {
std::cout << "output: " << data[i] << std::endl; std::cout << "output: " << data[i] << std::endl;
} }
} }
} // namespace paddle } // namespace paddle
int main() { int main() {
// Download address of the model http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/mobilenet.tar.gz // Download address of the model http://paddle-inference-dist.cdn.bcebos.com/tensorrt_test/mobilenet.tar.gz
paddle::RunTensorRT(1, "./mobilenet"); paddle::RunTensorRT(1, "./mobilenet");
return 0; return 0;
...@@ -120,11 +122,11 @@ The parameters of the neural network are redundant to some extent. In many tasks ...@@ -120,11 +122,11 @@ The parameters of the neural network are redundant to some extent. In many tasks
```shell ```shell
cd SAMPLE_BASE_DIR/sample cd SAMPLE_BASE_DIR/sample
# sh run_impl.sh {the address of inference libraries} {the name of test script} {model directories} # sh run_impl.sh {the address of inference libraries} {the name of test script} {model directories}
# We generate 500 input data to simulate the process, and it's suggested that you use real example for experiment. # We generate 500 input data to simulate the process, and it's suggested that you use real example for experiment.
sh run_impl.sh BASE_DIR/fluid_inference_install_dir/ fluid_generate_calib_test SAMPLE_BASE_DIR/sample/mobilenetv1 sh run_impl.sh BASE_DIR/fluid_inference_install_dir/ fluid_generate_calib_test SAMPLE_BASE_DIR/sample/mobilenetv1
``` ```
After the running period, there will be a new file named trt_calib_* under the `SAMPLE_BASE_DIR/sample/build/mobilenetv1` model directory, which is the calibration table. After the running period, there will be a new file named trt_calib_* under the `SAMPLE_BASE_DIR/sample/build/mobilenetv1` model directory, which is the calibration table.
``` shell ``` shell
...@@ -137,8 +139,8 @@ The parameters of the neural network are redundant to some extent. In many tasks ...@@ -137,8 +139,8 @@ The parameters of the neural network are redundant to some extent. In many tasks
## <a name="Paddle-TRT subgraph operation principle">Paddle-TRT subgraph operation principle</a> ## <a name="Paddle-TRT subgraph operation principle">Paddle-TRT subgraph operation principle</a>
Subgraph is used to integrate TensorRT in PaddlePaddle. After model is loaded, neural network can be represented as a computing graph composed of variables and computing nodes. Functions Paddle TensorRT implements are to scan the whole picture, discover subgraphs that can be optimized with TensorRT and replace them with TensorRT nodes. During the inference of model, Paddle will call TensorRT library to optimize TensorRT nodes and call native library of Paddle to optimize other nodes. During the inference, TensorRT can integrate Op horizonally and vertically to filter redundant Ops and is able to choose appropriate kernel for specific Op in specific platform to speed up the inference of model. Subgraph is used to integrate TensorRT in PaddlePaddle. After model is loaded, neural network can be represented as a computing graph composed of variables and computing nodes. Functions Paddle TensorRT implements are to scan the whole picture, discover subgraphs that can be optimized with TensorRT and replace them with TensorRT nodes. During the inference of model, Paddle will call TensorRT library to optimize TensorRT nodes and call native library of Paddle to optimize other nodes. During the inference, TensorRT can integrate Op horizonally and vertically to filter redundant Ops and is able to choose appropriate kernel for specific Op in specific platform to speed up the inference of model.
A simple model expresses the process : A simple model expresses the process :
**Original Network** **Original Network**
<p align="center"> <p align="center">
...@@ -151,5 +153,3 @@ A simple model expresses the process : ...@@ -151,5 +153,3 @@ A simple model expresses the process :
</p> </p>
We can see in the Original Network that the green nodes represent nodes supported by TensorRT, the red nodes represent variables in network and yellow nodes represent nodes which can only be operated by native functions in Paddle. Green nodes in original network are extracted to compose subgraph which is replaced by a single TensorRT node to be transformed into `block-25` node in network. When such nodes are encountered during the runtime, TensorRT library will be called to execute them. We can see in the Original Network that the green nodes represent nodes supported by TensorRT, the red nodes represent variables in network and yellow nodes represent nodes which can only be operated by native functions in Paddle. Green nodes in original network are extracted to compose subgraph which is replaced by a single TensorRT node to be transformed into `block-25` node in network. When such nodes are encountered during the runtime, TensorRT library will be called to execute them.
...@@ -62,7 +62,7 @@ BOOL类型。如果设置为True, GPU操作中的一些锁将被释放,Paralle ...@@ -62,7 +62,7 @@ BOOL类型。如果设置为True, GPU操作中的一些锁将被释放,Paralle
类型为bool,sync_batch_norm表示是否使用同步的批正则化,即在训练阶段通过多个设备同步均值和方差。 类型为bool,sync_batch_norm表示是否使用同步的批正则化,即在训练阶段通过多个设备同步均值和方差。
当前的实现不支持FP16培训和CPU。仅在一台机器上进行同步式批正则,不适用于多台机器。 当前的实现不支持FP16训练和CPU。仅在一台机器上进行同步式批正则,不适用于多台机器。
默认为 False。 默认为 False。
...@@ -1717,11 +1717,3 @@ WeightNormParamAttr ...@@ -1717,11 +1717,3 @@ WeightNormParamAttr
param_attr=WeightNormParamAttr( param_attr=WeightNormParamAttr(
dim=None, dim=None,
name='weight_norm_param')) name='weight_norm_param'))
...@@ -28,15 +28,18 @@ ...@@ -28,15 +28,18 @@
存储预测模型 存储预测模型
=========== ===========
存储预测模型时,一般通过 :code:`fluid.io.save_inference_model` 接口对默认的 :code:`fluid.Program` 进行裁剪,只保留预测 :code:`predict_var` 所需部分。
裁剪后的 program 会保存在指定路径 ./infer_model/__model__ 下,参数会保存到 ./infer_model 下的各个独立文件。
示例代码如下:
.. code-block:: python .. code-block:: python
exe = fluid.Executor(fluid.CPUPlace()) exe = fluid.Executor(fluid.CPUPlace())
path = "./infer_model" path = "./infer_model"
fluid.io.save_inference_model(dirname=path, feeded_var_names=['img'], fluid.io.save_inference_model(dirname=path, feeded_var_names=['img'],
target_vars=[predict_var], executor=exe) target_vars=[predict_var], executor=exe)
在这个示例中,:code:`fluid.io.save_inference_model` 接口对默认的 :code:`fluid.Program` 进行裁剪,只保留预测 :code:`predict_var` 所需部分。
裁剪后的 :code:`program` 会保存在 :code:`./infer_model/__model__` 下,参数会保存到 :code:`./infer_model` 下的各个独立文件。
加载预测模型 加载预测模型
=========== ===========
...@@ -45,11 +48,11 @@ ...@@ -45,11 +48,11 @@
exe = fluid.Executor(fluid.CPUPlace()) exe = fluid.Executor(fluid.CPUPlace())
path = "./infer_model" path = "./infer_model"
[inference_program, feed_target_names, fetch_targets] = [inference_program, feed_target_names, fetch_targets] =
fluid.io.load_inference_model(dirname=path, executor=exe) fluid.io.load_inference_model(dirname=path, executor=exe)
results = exe.run(inference_program, results = exe.run(inference_program,
feed={feed_target_names[0]: tensor_img}, feed={feed_target_names[0]: tensor_img},
fetch_list=fetch_targets) fetch_list=fetch_targets)
在这个示例中,首先调用 :code:`fluid.io.load_inference_model` 接口,获得预测的 :code:`program` 、输入数据的 :code:`variable` 名称和输出结果的 :code:`variable` ; 在这个示例中,首先调用 :code:`fluid.io.load_inference_model` 接口,获得预测的 :code:`inference_program` 、输入数据的名称 :code:`feed_target_names` 和输出结果的 :code:`fetch_targets` ;
然后调用 :code:`executor` 执行预测的 :code:`program` 获得预测结果。 然后调用 :code:`executor` 执行预测的 :code:`inference_program` 获得预测结果。
...@@ -28,15 +28,18 @@ There are two formats of saved inference model, which are controlled by :code:`m ...@@ -28,15 +28,18 @@ There are two formats of saved inference model, which are controlled by :code:`m
Save Inference model Save Inference model
=============================== ===============================
To save an inference model, we normally use :code:`fluid.io.save_inference_model` to tailor the default :code:`fluid.Program` and only keep the parts useful for predicting :code:`predict_var`.
After being tailored, :code:`program` will be saved under :code:`./infer_model/__model__` while the parameters will be saved into independent files under :code:`./infer_model` .
Sample Code:
.. code-block:: python .. code-block:: python
exe = fluid.Executor(fluid.CPUPlace()) exe = fluid.Executor(fluid.CPUPlace())
path = "./infer_model" path = "./infer_model"
fluid.io.save_inference_model(dirname=path, feeded_var_names=['img'], fluid.io.save_inference_model(dirname=path, feeded_var_names=['img'],
target_vars=[predict_var], executor=exe) target_vars=[predict_var], executor=exe)
In this example, :code:`fluid.io.save_inference_model` will tailor default :code:`fluid.Program` into useful parts for predicting :code:`predict_var` .
After being tailored, :code:`program` will be saved under :code:`./infer_model/__model__` while parameters will be saved into independent files under :code:`./infer_model` .
Load Inference Model Load Inference Model
===================== =====================
...@@ -45,11 +48,11 @@ Load Inference Model ...@@ -45,11 +48,11 @@ Load Inference Model
exe = fluid.Executor(fluid.CPUPlace()) exe = fluid.Executor(fluid.CPUPlace())
path = "./infer_model" path = "./infer_model"
[inference_program, feed_target_names, fetch_targets] = [inference_program, feed_target_names, fetch_targets] =
fluid.io.load_inference_model(dirname=path, executor=exe) fluid.io.load_inference_model(dirname=path, executor=exe)
results = exe.run(inference_program, results = exe.run(inference_program,
feed={feed_target_names[0]: tensor_img}, feed={feed_target_names[0]: tensor_img},
fetch_list=fetch_targets) fetch_list=fetch_targets)
In this example, at first we call :code:`fluid.io.load_inference_model` to get inference :code:`program` , :code:`variable` name of input data and :code:`variable` of output; In this example, at first we call :code:`fluid.io.load_inference_model` to get inference :code:`inference_program` , :code:`feed_target_names`-name of input data and :code:`fetch_targets` of output;
then call :code:`executor` to run inference :code:`program` to get inferred result. then call :code:`executor` to run inference :code:`inference_program` to get inferred result.
\ No newline at end of file
*** # **Installation on Windows**
# **Install on Windows** ## Operating Environment
This instruction will show you how to install PaddlePaddle on Windows. The following conditions must be met before you begin to install: * *64-bit operating system*
* *Windows 7/8, Windows 10 Pro/Enterprise*
* *Python 2.7/3.5/3.6/3.7*
* *pip or pip3 >= 9.0.1*
* *a 64-bit desktop or laptop* ### Precautions
* *Windows 7/8 , Windows 10 Professional/Enterprise Edition*
**Note** : * The default installation package requires your computer to support AVX instruction set and MKL. If your environment doesn’t support AVX instruction set and MKL, please download [these](./Tables.html/#ciwhls-release) `no-avx`, `openblas` versions of installation package.
* The current version doesn’t support functions related to NCCL and distributed learning.
* The current version does not support NCCL, distributed training related functions. ## CPU or GPU
* If your computer doesn’t have NVIDIA® GPU, please install the CPU version of PaddlePaddle
* If your computer has NVIDIA® GPU, and it satisfies the following requirements, we recommend you to install the GPU version of PaddlePaddle
* *CUDA Toolkit 8.0 with cuDNN v7*
* *GPU's computing capability exceeds 1.0*
Please refer to the NVIDIA official documents for the installation process and the configuration methods of [CUDA](https://docs.nvidia.com/cuda/cuda-installation-guide-linux/) and [cuDNN](https://docs.nvidia.com/deeplearning/sdk/cudnn-install/).
## Installation Method
## Installation Steps There are 3 ways to install PaddlePaddle on Windows:
### ***Install through pip*** * pip installation (recommended)
* [Docker installation](./install_Docker.html)
* [source code compilation and installation](./compile/compile_Windows.html/#win_source)
* Check your Python versions We would like to introduce the pip installation here.
Python2.7.15,Python3.5.x,Python3.6.x,Python3.7.x on [Official Python](https://www.python.org/downloads/) are supported. ## Installation Process
* Check your pip version
Version of pip or pip3 should be equal to or above 9.0.1 . * CPU version of PaddlePaddle: `pip install paddlepaddle` or `pip3 install paddlepaddle`
* GPU version of PaddlePaddle: `pip install paddlepaddle-gpu` or `pip3 install paddlepaddle-gpu`
* Install PaddlePaddle There is a checking function below for [verifyig whether the installation is successful](#check). If you have any further questions, please check the [FAQ part](./FAQ.html).
* ***CPU version of PaddlePaddle***: Notice:
Execute `pip install paddlepaddle` or `pip3 install paddlepaddle` to download and install PaddlePaddle.
* ***GPU version of PaddlePaddle***: * The version of pip and the version of python should be corresponding: python2.7 corresponds to `pip`; python3.x corresponds to `pip3`.
Execute `pip install paddlepaddle-gpu`(python2.7) or `pip3 install paddlepaddle-gpu`(python3.x) to download and install PaddlePaddle. * `pip install paddlepaddle-gpu` This command will install PaddlePaddle that supports CUDA 8.0 cuDNN v7. Currently, PaddlePaddle doesn't support any other version of CUDA or cuDNN on Windows.
## ***Verify installation***
After completing the installation, you can use `python` or `python3` to enter the python interpreter and then use `import paddle.fluid` to verify that the installation was successful. <a name="check"></a>
## Installation Verification
After completing the installation process, you can use `python` or `python3` to enter python interpreter and input `import paddle.fluid as fluid` and then `fluid.install_check.run_check()` to check whether the installation is successful.
## ***How to uninstall*** If you see `Your Paddle Fluid is installed succesfully!`, your installation is verified successful.
* ***CPU version of PaddlePaddle***: ## Uninstall PaddlePaddle
Use the following command to uninstall PaddlePaddle : `pip uninstallpaddlepaddle `or `pip3 uninstall paddlepaddle`
* ***GPU version of PaddlePaddle***: * ***CPU version of PaddlePaddle***: `pip uninstall paddlepaddle` or `pip3 uninstall paddlepaddle`
Use the following command to uninstall PaddlePaddle : `pip uninstall paddlepaddle-gpu` or `pip3 uninstall paddlepaddle-gpu`
* ***GPU version of PaddlePaddle***: `pip uninstall paddlepaddle-gpu` or `pip3 uninstall paddlepaddle-gpu`
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册