README.md 6.8 KB
Newer Older
L
LDOUBLEV 已提交
1 2 3 4
# OCR Pipeline WebService

(English|[简体中文](./README_CN.md))

L
LDOUBLEV 已提交
5 6 7
PaddleOCR provides 2 service deployment methods:
- Based on **PaddleHub Serving**: Code path is "`./deploy/hubserving`". Please refer to the [tutorial](../../deploy/hubserving/readme_en.md)
- Based on **PaddleServing**: Code path is "`./deploy/pdserving`". Please follow this tutorial.
L
LDOUBLEV 已提交
8

L
LDOUBLEV 已提交
9 10 11 12 13 14 15 16 17 18
# Service deployment based on PaddleServing  

This document will introduce how to use the [PaddleServing](https://github.com/PaddlePaddle/Serving/blob/develop/README.md) to deploy the PPOCR dynamic graph model as a pipeline online service.

Some Key Features of Paddle Serving:
- Integrate with Paddle training pipeline seamlessly, most paddle models can be deployed with one line command.
- Industrial serving features supported, such as models management, online loading, online A/B testing etc.
- Highly concurrent and efficient communication between clients and servers supported.

The introduction and tutorial of Paddle Serving service deployment framework reference [document](https://github.com/PaddlePaddle/Serving/blob/develop/README.md).
L
LDOUBLEV 已提交
19 20 21


## Contents
L
LDOUBLEV 已提交
22 23 24 25
- [Environmental preparation](#environmental-preparation)
- [Model conversion](#model-conversion)
- [Paddle Serving pipeline deployment](#paddle-serving-pipeline-deployment)
- [FAQ](#faq)
L
LDOUBLEV 已提交
26

L
LDOUBLEV 已提交
27
<a name="environmental-preparation"></a>
L
LDOUBLEV 已提交
28 29 30 31
## Environmental preparation

Need to prepare PaddleOCR operating environment and Paddle Serving operating environment.

L
LDOUBLEV 已提交
32
1. Prepare PaddleOCR operating environment reference [link](../../doc/doc_ch/installation.md).
L
LDOUBLEV 已提交
33 34 35

2. Prepare the operating environment of PaddleServing, the steps are as follows

L
LDOUBLEV 已提交
36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71
    Install serving which used to start the service
    ```
    pip3 install paddle-serving-server==0.5.0 # for CPU
    pip3 install paddle-serving-server-gpu==0.5.0 # for GPU
    # Other GPU environments need to confirm the environment and then choose to execute the following commands
    pip3 install paddle-serving-server-gpu==0.5.0.post9 # GPU with CUDA9.0
    pip3 install paddle-serving-server-gpu==0.5.0.post10 # GPU with CUDA10.0
    pip3 install paddle-serving-server-gpu==0.5.0.post101 # GPU with CUDA10.1 + TensorRT6
    pip3 install paddle-serving-server-gpu==0.5.0.post11 # GPU with CUDA10.1 + TensorRT7
    ```

3. Install the client to send requests to the service
    ```
    pip3 install paddle-serving-client==0.5.0 # for CPU

    pip3 install paddle-serving-client-gpu==0.5.0 # for GPU
    ```

4. Install serving-app
    ```
    pip3 install paddle-serving-app==0.3.0
    # fix local_predict to support load dynamic model
    # find the install directoory of paddle_serving_app
    vim /usr/local/lib/python3.7/site-packages/paddle_serving_app/local_predict.py
    # replace line 85 of local_predict.py config = AnalysisConfig(model_path) with:
    if os.path.exists(os.path.join(model_path, "__params__")):
        config = AnalysisConfig(os.path.join(model_path, "__model__"), os.path.join(model_path, "__params__"))
    else:
        config = AnalysisConfig(model_path)
    ```


   **note:** If you want to install the latest version of PaddleServing, refer to [link](https://github.com/PaddlePaddle/Serving/blob/develop/doc/LATEST_PACKAGES.md).


<a name="model-conversion"></a>
L
LDOUBLEV 已提交
72 73 74
## Model conversion
When using PaddleServing for service deployment, you need to convert the saved inference model into a serving model that is easy to deploy.

L
LDOUBLEV 已提交
75
Firstly, download the [inference model](https://github.com/PaddlePaddle/PaddleOCR#pp-ocr-20-series-model-listupdate-on-dec-15) of PPOCR
L
LDOUBLEV 已提交
76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112
```
# Download and unzip the OCR text detection model
wget https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_server_v2.0_det_infer.tar && tar xf ch_ppocr_server_v2.0_det_infer.tar
# Download and unzip the OCR text recognition model
wget https://paddleocr.bj.bcebos.com/dygraph_v2.0/ch/ch_ppocr_server_v2.0_rec_infer.tar && tar xf ch_ppocr_server_v2.0_rec_infer.tar

# Conversion detection model
python3 -m paddle_serving_client.convert --dirname ./ch_ppocr_server_v2.0_det_infer/ \
                                         --model_filename inference.pdmodel          \
                                         --params_filename inference.pdiparams       \
                                         --serving_server ./ppocr_det_server_2.0_serving/ \
                                         --serving_client ./ppocr_det_server_2.0_client/

# Conversion recognition model
python3 -m paddle_serving_client.convert --dirname ./ch_ppocr_server_v2.0_rec_infer/ \
                                         --model_filename inference.pdmodel          \
                                         --params_filename inference.pdiparams       \
                                         --serving_server ./ppocr_rec_server_2.0_serving/  \
                                         --serving_client ./ppocr_rec_server_2.0_client/

```

After the detection model is converted, there will be additional folders of `ppocr_det_server_2.0_serving` and `ppocr_det_server_2.0_client` in the current folder, with the following format:
```
|- ppocr_det_server_2.0_serving/
   |- __model__
   |- __params__
   |- serving_server_conf.prototxt
   |- serving_server_conf.stream.prototxt

|- ppocr_det_server_2.0_client
   |- serving_client_conf.prototxt
   |- serving_client_conf.stream.prototxt

```
The recognition model is the same.

L
LDOUBLEV 已提交
113
<a name="paddle-serving-pipeline-deployment"></a>
L
LDOUBLEV 已提交
114 115 116
## Paddle Serving pipeline deployment

1. Download the PaddleOCR code, if you have already downloaded it, you can skip this step.
L
LDOUBLEV 已提交
117 118 119 120 121 122 123 124 125 126 127 128 129 130 131
    ```
    git clone https://github.com/PaddlePaddle/PaddleOCR

    # Enter the working directory  
    cd PaddleOCR/deploy/pdserver/
    ```

    The pdserver directory contains the code to start the pipeline service and send prediction requests, including:
    ```
    __init__.py
    config.yml # Start the service configuration file
    ocr_reader.py # OCR model pre-processing and post-processing code implementation
    pipeline_http_client.py # Script to send pipeline prediction request
    web_service.py # Start the script of the pipeline server
    ```
L
LDOUBLEV 已提交
132 133

2. Run the following command to start the service.
L
LDOUBLEV 已提交
134 135 136 137 138 139
    ```
    # Start the service and save the running log in log.txt
    python3 web_service.py &>log.txt &
    ```
    After the service is successfully started, a log similar to the following will be printed in log.txt
    ![](./imgs/start_server.png)
L
LDOUBLEV 已提交
140 141

3. Send service request
L
LDOUBLEV 已提交
142 143 144 145 146
    ```
    python3 pipeline_http_client.py
    ```
    After successfully running, the predicted result of the model will be printed in the cmd window. An example of the result is:
    ![](./imgs/results.png)  
L
LDOUBLEV 已提交
147

L
LDOUBLEV 已提交
148
<a name="faq"></a>
L
LDOUBLEV 已提交
149
## FAQ
L
LDOUBLEV 已提交
150 151
** Q1**: No result return after sending the request.

L
LDOUBLEV 已提交
152 153 154 155 156
** A1**: Do not set the proxy when starting the service and sending the request. You can close the proxy before starting the service and before sending the request. The command to close the proxy is:
```
unset https_proxy
unset http_proxy
```