diff --git a/python/examples/bert/test_multi_fetch_client.py b/python/examples/bert/test_multi_fetch_client.py index 3749e4d384498a858c6cc784c3bfc387d7fd791f..c15c4d4deaf282c432ff0990ee03c6e80daeee74 100644 --- a/python/examples/bert/test_multi_fetch_client.py +++ b/python/examples/bert/test_multi_fetch_client.py @@ -17,8 +17,7 @@ from paddle_serving_app.reader import ChineseBertReader import sys client = Client() -client.load_client_config( - "./bert_multi_fetch_client/serving_client_conf.prototxt") +client.load_client_config("./bert_seq32_client/serving_client_conf.prototxt") client.connect(["127.0.0.1:9292"]) reader = ChineseBertReader({"max_seq_len": 32}) diff --git a/tools/serving_build.sh b/tools/serving_build.sh index 0f1a8cb2e924e94f7d2762f743688e9a8b7de1a8..8e78e13ef8e86b55af6a90df1b9235611508c0ba 100644 --- a/tools/serving_build.sh +++ b/tools/serving_build.sh @@ -341,7 +341,7 @@ function python_test_multi_fetch() { #download model (max_seq_len=32) wget https://paddle-serving.bj.bcebos.com/bert_example/bert_multi_fetch.tar.gz tar -xzvf bert_multi_fetch.tar.gz - check_cmd "python -m paddle_serving_server.serve --model bert_multi_fetch_model --port 9292 &" + check_cmd "python -m paddle_serving_server.serve --model bert_seq32_model --port 9292 &" sleep 5 check_cmd "head -n 8 data-c.txt | python test_multi_fetch_client.py" kill_server_process @@ -351,7 +351,7 @@ function python_test_multi_fetch() { #download model (max_seq_len=32) wget https://paddle-serving.bj.bcebos.com/bert_example/bert_multi_fetch.tar.gz tar -xzvf bert_multi_fetch.tar.gz - check_cmd "python -m paddle_serving_server_gpu.serve --model bert_multi_fetch_model --port 9292 --gpu_ids 0 &" + check_cmd "python -m paddle_serving_server_gpu.serve --model bert_seq32_model --port 9292 --gpu_ids 0 &" sleep 5 check_cmd "head -n 8 data-c.txt | python test_multi_fetch_client.py" kill_server_process