diff --git a/python/examples/criteo_ctr/benchmark_batch.py b/python/examples/criteo_ctr/benchmark_batch.py index 47b63a6ade0c21bdc82a5c67d65b39ffc614e06c..1e4348c99dc0d960b1818ea6f0eb1ae2f5bd2ccb 100644 --- a/python/examples/criteo_ctr/benchmark_batch.py +++ b/python/examples/criteo_ctr/benchmark_batch.py @@ -55,8 +55,7 @@ def single_func(idx, resource): for i in range(1, 27): feed_dict["sparse_{}".format(i - 1)] = data[0][i] feed_batch.append(feed_dict) - result = client.batch_predict( - feed_batch=feed_batch, fetch=fetch) + result = client.predict(feed=feed_batch, fetch=fetch) else: print("unsupport batch size {}".format(args.batch_size)) diff --git a/python/examples/criteo_ctr_with_cube/benchmark_batch.py b/python/examples/criteo_ctr_with_cube/benchmark_batch.py index b4b15892375e830486afa320151fac619aab2ba7..f395d327ea8e3470db08a7f472b6b6ec5a82daf1 100755 --- a/python/examples/criteo_ctr_with_cube/benchmark_batch.py +++ b/python/examples/criteo_ctr_with_cube/benchmark_batch.py @@ -56,8 +56,7 @@ def single_func(idx, resource): feed_dict["embedding_{}.tmp_0".format(i - 1)] = data[0][ i] feed_batch.append(feed_dict) - result = client.batch_predict( - feed_batch=feed_batch, fetch=fetch) + result = client.batch_predict(feed=feed_batch, fetch=fetch) else: print("unsupport batch size {}".format(args.batch_size)) diff --git a/python/examples/imagenet/benchmark_batch.py b/python/examples/imagenet/benchmark_batch.py index 7477100971d1a098bb075c94f75490c16b53b862..e531425770cbf9102b7ebd2f5b082c5c4aa14e71 100644 --- a/python/examples/imagenet/benchmark_batch.py +++ b/python/examples/imagenet/benchmark_batch.py @@ -50,8 +50,7 @@ def single_func(idx, resource): img = reader.process_image(img_list[i]) img = img.reshape(-1) feed_batch.append({"image": img}) - result = client.batch_predict( - feed_batch=feed_batch, fetch=fetch) + result = client.predict(feed=feed_batch, fetch=fetch) else: print("unsupport batch size {}".format(args.batch_size)) diff --git a/python/examples/imdb/benchmark_batch.py b/python/examples/imdb/benchmark_batch.py index 302d63352ca20bf7e455ad1a66ead22f63dbe846..d36704a7631e963fd51220aa3c3d9a350515ebfd 100644 --- a/python/examples/imdb/benchmark_batch.py +++ b/python/examples/imdb/benchmark_batch.py @@ -42,8 +42,7 @@ def single_func(idx, resource): for bi in range(args.batch_size): word_ids, label = imdb_dataset.get_words_and_label(line) feed_batch.append({"words": word_ids}) - result = client.batch_predict( - feed_batch=feed_batch, fetch=["prediction"]) + result = client.predict(feed=feed_batch, fetch=["prediction"]) else: print("unsupport batch size {}".format(args.batch_size))