CUBE_LOCAL.md 5.5 KB
Newer Older
J
Jiawei Wang 已提交
1 2 3 4 5 6 7 8 9 10
# Cube: Sparse Parameter Server (Local Mode)

## Overview


There are two examples on CTR under python / examples, they are criteo_ctr, criteo_ctr_with_cube. The former is to save the entire model during training, including sparse parameters. The latter is to cut out the sparse parameters and save them into two parts, one is the sparse parameter and the other is the dense parameter. Because the scale of sparse parameters is very large in industrial cases, reaching the order of 10 ^ 9. Therefore, it is not practical to start large-scale sparse parameter prediction on one machine. Therefore, we introduced Baidu's industrial-grade product Cube on the sparse parameter server for many years to provide distributed sparse parameter services.

The local mode of Cube is different from distributed Cube, which is designed to be convenient for developers to use in experiments and demos. If there is a demand for distributed sparse parameter service, please continue reading [Distributed Cube Usage Guide] (Distributed Cube) after reading this document (still developing).


J
Jiawei Wang 已提交
11
## Example
J
Jiawei Wang 已提交
12
in directory python/example/criteo_ctr_with_cube, run
J
Jiawei Wang 已提交
13

J
Jiawei Wang 已提交
14
```
J
Jiawei Wang 已提交
15
python local_train.py # train model
J
Jiawei Wang 已提交
16 17 18 19
cp ../../../build_server/core/predictor/seq_generator seq_generator # copy Sequence File generator
cp ../../../build_server/output/bin/cube* ./cube/ # copy Cube tool kits
cp ../../../build_server/core/cube/cube-api/cube-cli ./cube/ # copy Cube Client
cube_prepare.sh & # start deliver script
J
Jiawei Wang 已提交
20 21 22
```
you will convert the Sparse Parameters from trained model to the Cube Server.

J
Jiawei Wang 已提交
23
## Components of Cube
J
Jiawei Wang 已提交
24

J
Jiawei Wang 已提交
25
### cube-builder
J
Jiawei Wang 已提交
26 27 28

cube-builder is a tool for generating model shard files and version management. As the cube is used for distributed sparse parameter services, for each node in the distribution, different shards need to be loaded. However, the generated sparse parameter file is often a large file, and it needs to be divided into different shards by a hash function. At the same time, industrial-level scenarios need to support regular model distribution and streaming training, so it is very important for the version management of the model. This is also the missing part when training and saving the model. Therefore, while the cube-builder generates the shards, You can also manually specify version information.

J
Jiawei Wang 已提交
29
### cube-server
J
Jiawei Wang 已提交
30 31 32

The cube-server is the sparse parameter server itself. It provides high-performance distributed query service through brpc, and makes remote calls through RestAPI.

J
Jiawei Wang 已提交
33
### cube-cli
J
Jiawei Wang 已提交
34 35 36 37

cube-cli is the client of cube-server. This part has been integrated into paddle serving. When we prepare the cube.conf configuration file and specify the kv_infer related op in the code of paddle serving server, cube-cli will Ready on the serving side.


J
Jiawei Wang 已提交
38 39 40 41 42 43 44 45 46 47 48 49
## Serving the Model Step by Step
### precondition

we need a trained model, and copy the tool kits from build_server folder.
```
python local_train.py # train model
cp ../../../build_server/core/predictor/seq_generator seq_generator  # copy Sequence File generator
cp ../../../build_server/output/bin/cube* ./cube/ # copy Cube tool kits
cp ../../../build_server/core/cube/cube-api/cube-cli ./cube/ # copy Cube Client
```

### Generate Sequence File from Sparse Parameter
J
Jiawei Wang 已提交
50 51 52 53 54 55 56

In order to get the model parameters from the training end to the prediction end, we need to convert the trained model from the Paddle model save format to the Sequence File format.

**why is Sequence File?**
Sequence File is a common format for the Hadoop File System. It was mentioned at the beginning of the article that distributed cubes can provide support for ultra-large-scale sparse parameter services, and large-scale sparse parameters are stored in distributed file systems in actual production environments. Hadoop File System is one of the most stable distributed open source. So the Sequence File format became the file format for the Cube loading model.

```
J
Jiawei Wang 已提交
57 58 59
mkdir -p cube_model
mkdir -p cube/data
./seq_generator ctr_serving_model/SparseFeatFactors ./cube_model/feature
J
Jiawei Wang 已提交
60 61
```

J
Jiawei Wang 已提交
62
### Generating Shards
J
Jiawei Wang 已提交
63 64 65 66

For the local version of Cube, the number of shard is 1. run

```
J
Jiawei Wang 已提交
67
cube-builder -dict_name=test_dict -job_mode=base -last_version=0 -cur_version=0 -depend_version=0 -input_path=./cube_model -output_path=./cube/data -shard_num=1  -only_build=false
J
Jiawei Wang 已提交
68 69
```

J
Jiawei Wang 已提交
70

J
Jiawei Wang 已提交
71
### Deliver to Cube-Server
J
Jiawei Wang 已提交
72 73 74

The process of the cube local version is very simple, you only need to store the index files. in ./data folder where the cube binary program is located.
```
J
Jiawei Wang 已提交
75 76
mv ./cube/data/0_0/test_dict_part0/* ./cube/data/
cd cube && ./cube &
J
Jiawei Wang 已提交
77 78 79
```


J
Jiawei Wang 已提交
80
### Cube-Client Verification
J
Jiawei Wang 已提交
81 82 83 84 85

this step is not necessary, but it can help you to verify if the model is ready.
```
./cube-cli -dict_name=test_dict -keys  keys -conf ./cube/cube.conf
```
J
Jiawei Wang 已提交
86 87 88 89
if you succeed, you will see this
<p align="center">
    <img src="doc/cube_cli.png" width="700">
</p>
J
Jiawei Wang 已提交
90 91

If you see that each key has a corresponding value output, it means that the delivery was successful. This file can also be used by Serving to perform cube query in general kv infer op in Serving.
J
Jiawei Wang 已提交
92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112


## Appendix: Configuration
the config file is cube.config located in python/examples/criteo_ctr_with_cube/cube/conf, this file is used by cube-cli.the Cube Local Mode users do not need to understand that just use it, it would be quite important in Cube Distributed Mode.

```
[{
    "dict_name": "test_dict",  //table name
    "shard": 1,  //shard num
    "dup": 1,  //duplicates
    "timeout": 200,
    "retry": 3,
    "backup_request": 100,
    "type": "ipport_list",
    "load_balancer": "rr",
    "nodes": [{
        "ipport_list": "list://127.0.0.1:8027" //IP list
    }]
}]
```