## CTR Prediction Service ([简体中文](./README_CN.md)|English) ### download criteo dataset ``` sh get_data.sh ``` ### download inference model ``` wget https://paddle-serving.bj.bcebos.com/criteo_ctr_example/criteo_ctr_demo_model.tar.gz tar xf criteo_ctr_demo_model.tar.gz mv models/ctr_client_conf . mv models/ctr_serving_model . ``` the directories like serving_server_model and serving_client_config will appear. ### Start RPC Inference Service ``` python -m paddle_serving_server.serve --model ctr_serving_model/ --port 9292 #CPU RPC Service python -m paddle_serving_server_gpu.serve --model ctr_serving_model/ --port 9292 --gpu_ids 0 #RPC Service on GPU 0 ``` ### RPC Infer ``` python test_client.py ctr_client_conf/serving_client_conf.prototxt raw_data/ ``` the latency will display in the end.