diff --git a/python/paddle_serving_app/README.md b/python/paddle_serving_app/README.md index 4cc9b4682e12946cf29b76fb4fe09e6f54a6f46d..1756b83993e67dcbc66b6809631c5e953eef08d7 100644 --- a/python/paddle_serving_app/README.md +++ b/python/paddle_serving_app/README.md @@ -1,4 +1,4 @@ -([简体中文](./README.md)|English) +([简体中文](./README_CN.md)|English) paddle_serving_app is a tool component of the Paddle Serving framework, and includes functions such as pre-training model download and data pre-processing methods. It is convenient for users to quickly test and deploy model examples, analyze the performance of prediction services, and debug model prediction services. @@ -142,7 +142,7 @@ This tool is convenient to analyze the proportion of time occupancy in the predi Load the trace file generated in the previous step through the load button, you can Visualize the time information of each stage of the forecast service. -As shown in next figure, the figure shows the timeline of GPU prediction service using [bert example] (https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/bert). +As shown in next figure, the figure shows the timeline of GPU prediction service using [bert example](https://github.com/PaddlePaddle/Serving/tree/develop/python/examples/bert). The server side starts service with 4 GPU cards, the client side starts 4 processes to request, and the batch size is 1. In the figure, bert_pre represents the data pre-processing stage of the client, and client_infer represents the stage where the client completes the sending of the prediction request to the receiving result. The process in the figure represents the process number of the client, and the second line of each process shows the timeline of each op of the server. @@ -155,7 +155,7 @@ The inference op of Paddle Serving is implemented based on Paddle inference lib. Before deploying the prediction service, you may need to check the input and output of the prediction service or check the resource consumption. Therefore, a local prediction tool is built into the paddle_serving_app, which is used in the same way as sending a request to the server through the client. -Taking [fit_a_line prediction service] (../examples/fit_a_line) as an example, the following code can be used to run local prediction. +Taking [fit_a_line prediction service](../examples/fit_a_line) as an example, the following code can be used to run local prediction. ```python from paddle_serving_app import Debugger diff --git a/python/paddle_serving_app/README_CN.md b/python/paddle_serving_app/README_CN.md index 05854556f9428b23c9d2217417f991b271881d06..75dcf9ae78bec0c00b7662f7427d3816feaeca3d 100644 --- a/python/paddle_serving_app/README_CN.md +++ b/python/paddle_serving_app/README_CN.md @@ -1,4 +1,4 @@ -(简体中文|[English](./README_CN.md)) +(简体中文|[English](./README.md)) paddle_serving_app是Paddle Serving框架的工具组件,包含了预训练模型下载、数据预处理方法等功能。方便用户快速体验和部署模型示例、分析预测服务性能、调试模型预测服务等。