RUN_IN_DOCKER.md 6.3 KB
Newer Older
1 2
# How to run PaddleServing in Docker

B
barrierye 已提交
3
([简体中文](RUN_IN_DOCKER_CN.md)|English)
B
fix doc  
barrierye 已提交
4

5 6 7 8
## Requirements

Docker (GPU version requires nvidia-docker to be installed on the GPU machine)

B
barrierye 已提交
9 10
This document takes Python2 as an example to show how to run Paddle Serving in docker. You can also use Python3 to run related commands by replacing `python` with `python3`.

11 12 13 14 15 16 17 18 19
## CPU

### Get docker image

You can get images in two ways:

1. Pull image directly

   ```bash
B
barrierye 已提交
20
   docker pull hub.baidubce.com/paddlepaddle/serving:latest
21 22 23 24
   ```

2. Building image based on dockerfile

B
barrierye 已提交
25
   Create a new folder and copy [Dockerfile](../tools/Dockerfile) to this folder, and run the following command:
26 27

   ```bash
B
barrierye 已提交
28
   docker build -t hub.baidubce.com/paddlepaddle/serving:latest .
29 30 31 32 33
   ```

### Create container

```bash
B
barrierye 已提交
34
docker run -p 9292:9292 --name test -dit hub.baidubce.com/paddlepaddle/serving:latest
35 36 37 38 39 40 41
docker exec -it test bash
```

The `-p` option is to map the `9292` port of the container to the `9292` port of the host.

### Install PaddleServing

42
In order to make the image smaller, the PaddleServing package is not installed in the image. You can run the following command to install it:
43 44 45 46 47

```bash
pip install paddle-serving-server
```

48 49 50 51 52 53
You may need to use a domestic mirror source (in China, you can use the Tsinghua mirror source of the following example) to speed up the download:

```shell
pip install paddle-serving-server -i https://pypi.tuna.tsinghua.edu.cn/simple
```

54 55 56 57 58 59 60 61 62 63 64 65 66 67
### Test example

Get the trained Boston house price prediction model by the following command:

```bash
wget --no-check-certificate https://paddle-serving.bj.bcebos.com/uci_housing.tar.gz
tar -xzf uci_housing.tar.gz
```

- Test HTTP service

  Running on the Server side (inside the container):

  ```bash
B
barrierye 已提交
68
  python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --port 9292 --name uci >std.log 2>err.log &
69 70 71 72 73
  ```

  Running on the Client side (inside or outside the container):

  ```bash
B
barrierye 已提交
74
  curl -H "Content-Type:application/json" -X POST -d '{"feed":{"x": [0.0137, -0.1136, 0.2553, -0.0692, 0.0582, -0.0727, -0.1583, -0.0584, 0.6283, 0.4919, 0.1856, 0.0795, -0.0332]}, "fetch":["price"]}' http://127.0.0.1:9292/uci/prediction
75 76 77 78 79 80 81
  ```

- Test RPC service

  Running on the Server side (inside the container):

  ```bash
B
barrierye 已提交
82
  python -m paddle_serving_server.serve --model uci_housing_model --thread 10 --port 9292 >std.log 2>err.log &
83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111
  ```

  Running following Python code on the Client side (inside or outside the container, The `paddle-serving-client` package needs to be installed):

  ```bash
  from paddle_serving_client import Client
  
  client = Client()
  client.load_client_config("uci_housing_client/serving_client_conf.prototxt")
  client.connect(["127.0.0.1:9292"])
  data = [0.0137, -0.1136, 0.2553, -0.0692, 0.0582, -0.0727,
          -0.1583, -0.0584, 0.6283, 0.4919, 0.1856, 0.0795, -0.0332]
  fetch_map = client.predict(feed={"x": data}, fetch=["price"])
  print(fetch_map)
  ```

  

## GPU

The GPU version is basically the same as the CPU version, with only some differences in interface naming (GPU version requires nvidia-docker to be installed on the GPU machine).

### Get docker image

You can also get images in two ways:

1. Pull image directly

   ```bash
B
barrierye 已提交
112
   nvidia-docker pull hub.baidubce.com/paddlepaddle/serving:latest-gpu
113 114 115 116
   ```

2. Building image based on dockerfile

B
barrierye 已提交
117
   Create a new folder and copy [Dockerfile.gpu](../tools/Dockerfile.gpu) to this folder, and run the following command:
118 119

   ```bash
B
barrierye 已提交
120
   nvidia-docker build -t hub.baidubce.com/paddlepaddle/serving:latest-gpu .
121 122 123 124 125
   ```

### Create container

```bash
B
barrierye 已提交
126
nvidia-docker run -p 9292:9292 --name test -dit hub.baidubce.com/paddlepaddle/serving:latest-gpu
127 128 129 130 131 132 133 134 135 136 137 138 139
nvidia-docker exec -it test bash
```

The `-p` option is to map the `9292` port of the container to the `9292` port of the host.

### Install PaddleServing

In order to make the image smaller, the PaddleServing package is not installed in the image. You can run the following command to install it:

```bash
pip install paddle-serving-server-gpu
```

140 141 142
You may need to use a domestic mirror source (in China, you can use the Tsinghua mirror source of the following example) to speed up the download:

```shell
B
barrierye 已提交
143
pip install paddle-serving-server-gpu -i https://pypi.tuna.tsinghua.edu.cn/simple
144 145
```

146 147
### Test example

B
barrierye 已提交
148 149 150 151 152
When running the GPU Server, you need to set the GPUs used by the prediction service through the `--gpu_ids` option, and the CPU is used by default. An error will be reported when the value of `--gpu_ids` exceeds the environment variable `CUDA_VISIBLE_DEVICES`. The following example specifies to use a GPU with index 0:
```shell
export CUDA_VISIBLE_DEVICES=0,1
python -m paddle_serving_server_gpu.serve --model uci_housing_model --port 9292 --gpu_ids 0
```
B
barrierye 已提交
153 154


155 156 157 158 159 160 161 162 163 164 165 166
Get the trained Boston house price prediction model by the following command:

```bash
wget --no-check-certificate https://paddle-serving.bj.bcebos.com/uci_housing.tar.gz
tar -xzf uci_housing.tar.gz
```

- Test HTTP service

  Running on the Server side (inside the container):

  ```bash
B
barrierye 已提交
167
  python -m paddle_serving_server_gpu.serve --model uci_housing_model --thread 10 --port 9292 --name uci --gpu_ids 0
168 169 170 171 172
  ```

  Running on the Client side (inside or outside the container):

  ```bash
B
barrierye 已提交
173
  curl -H "Content-Type:application/json" -X POST -d '{"feed":{"x": [0.0137, -0.1136, 0.2553, -0.0692, 0.0582, -0.0727, -0.1583, -0.0584, 0.6283, 0.4919, 0.1856, 0.0795, -0.0332]}, "fetch":["price"]}' http://127.0.0.1:9292/uci/prediction
174 175 176 177 178 179 180
  ```

- Test RPC service

  Running on the Server side (inside the container):

  ```bash
B
barrierye 已提交
181
  python -m paddle_serving_server_gpu.serve --model uci_housing_model --thread 10 --port 9292 --gpu_ids 0
182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197
  ```

  Running following Python code on the Client side (inside or outside the container, The `paddle-serving-client` package needs to be installed):

  ```bash
  from paddle_serving_client import Client
  
  client = Client()
  client.load_client_config("uci_housing_client/serving_client_conf.prototxt")
  client.connect(["127.0.0.1:9292"])
  data = [0.0137, -0.1136, 0.2553, -0.0692, 0.0582, -0.0727,
          -0.1583, -0.0584, 0.6283, 0.4919, 0.1856, 0.0795, -0.0332]
  fetch_map = client.predict(feed={"x": data}, fetch=["price"])
  print(fetch_map)
  ```

B
barrierye 已提交
198 199 200 201 202



## Attention

B
fix doc  
barrierye 已提交
203
The images provided by this document are all runtime images, which do not support compilation. If you want to compile from source, refer to [COMPILE](COMPILE.md).