diff --git a/PaddleRec/ctr/infer.py b/PaddleRec/ctr/infer.py index 19f7013b305e3900fe05b7575cfc20c331ca5daf..2f622629ec487b076f382f141061804a760c71ac 100644 --- a/PaddleRec/ctr/infer.py +++ b/PaddleRec/ctr/infer.py @@ -11,9 +11,7 @@ import paddle.fluid as fluid import reader from network_conf import ctr_dnn_model - -logging.basicConfig( - format='%(asctime)s - %(levelname)s - %(message)s') +logging.basicConfig(format='%(asctime)s - %(levelname)s - %(message)s') logger = logging.getLogger("fluid") logger.setLevel(logging.INFO) @@ -53,22 +51,26 @@ def infer(): args = parse_args() place = fluid.CPUPlace() - inference_scope = fluid.core.Scope() - + inference_scope = fluid.Scope() + dataset = reader.CriteoDataset(args.sparse_feature_dim) - test_reader = paddle.batch(dataset.test([args.data_path]), batch_size=args.batch_size) - + test_reader = paddle.batch( + dataset.test([args.data_path]), batch_size=args.batch_size) + startup_program = fluid.framework.Program() test_program = fluid.framework.Program() with fluid.framework.program_guard(test_program, startup_program): - loss, auc_var, batch_auc_var, _, data_list = ctr_dnn_model(args.embedding_size, args.sparse_feature_dim, False) - + loss, auc_var, batch_auc_var, _, data_list = ctr_dnn_model( + args.embedding_size, args.sparse_feature_dim, False) + exe = fluid.Executor(place) feeder = fluid.DataFeeder(feed_list=data_list, place=place) - fluid.io.load_persistables(executor=exe, dirname=args.model_path, - main_program=fluid.default_main_program()) + fluid.io.load_persistables( + executor=exe, + dirname=args.model_path, + main_program=fluid.default_main_program()) def set_zero(var_name): param = inference_scope.var(var_name).get_tensor() @@ -81,10 +83,11 @@ def infer(): for batch_id, data in enumerate(test_reader()): loss_val, auc_val = exe.run(test_program, - feed=feeder.feed(data), - fetch_list=[loss, auc_var]) + feed=feeder.feed(data), + fetch_list=[loss, auc_var]) if batch_id % 100 == 0: - logger.info("TEST --> batch: {} loss: {} auc: {}".format(batch_id, loss_val/args.batch_size, auc_val)) + logger.info("TEST --> batch: {} loss: {} auc: {}".format( + batch_id, loss_val / args.batch_size, auc_val)) if __name__ == '__main__': diff --git a/PaddleRec/din/infer.py b/PaddleRec/din/infer.py index 1ca763a484676bbbc680b38a2ade5c79b161b853..fc1484e51ec4fc42b98748723d34b26e52946848 100644 --- a/PaddleRec/din/infer.py +++ b/PaddleRec/din/infer.py @@ -28,9 +28,15 @@ logger.setLevel(logging.INFO) def parse_args(): parser = argparse.ArgumentParser(description="PaddlePaddle DIN example") parser.add_argument( - '--model_path', type=str, required=True, help="path of model parameters") + '--model_path', + type=str, + required=True, + help="path of model parameters") parser.add_argument( - '--test_path', type=str, default='data/paddle_test.txt.bak', help='dir of test file') + '--test_path', + type=str, + default='data/paddle_test.txt.bak', + help='dir of test file') parser.add_argument( '--use_cuda', type=int, default=0, help='whether to use gpu') @@ -64,7 +70,7 @@ def infer(): data_reader, _ = reader.prepare_reader(args.test_path, 32 * 16) place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace() - inference_scope = fluid.core.Scope() + inference_scope = fluid.Scope() exe = fluid.Executor(place) diff --git a/PaddleRec/gru4rec/infer.py b/PaddleRec/gru4rec/infer.py index ec113260bb4391cb970c515a2c958ae71b6176b7..bc459c28a9b24761b202dc5d8110d583322abdeb 100644 --- a/PaddleRec/gru4rec/infer.py +++ b/PaddleRec/gru4rec/infer.py @@ -37,7 +37,7 @@ def infer(test_reader, use_cuda, model_path): place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace() exe = fluid.Executor(place) - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): infer_program, feed_target_names, fetch_vars = fluid.io.load_inference_model( model_path, exe) accum_num_recall = 0.0 diff --git a/PaddleRec/gru4rec/infer_sample_neg.py b/PaddleRec/gru4rec/infer_sample_neg.py index 8839d6ceee2959b81edc2e50ef3b7ae9049c1a1f..48458e82b4fe2bbc7141c3e45469b8414d87ece4 100644 --- a/PaddleRec/gru4rec/infer_sample_neg.py +++ b/PaddleRec/gru4rec/infer_sample_neg.py @@ -40,7 +40,7 @@ def infer(args, vocab_size, test_reader, use_cuda): exe = fluid.Executor(place) hid_size = args.hid_size batch_size = args.batch_size - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): main_program = fluid.Program() with fluid.program_guard(main_program): acc = net.infer_network(vocab_size, batch_size, hid_size) diff --git a/PaddleRec/multiview_simnet/infer.py b/PaddleRec/multiview_simnet/infer.py index 80262bb50695d9c064441bd7edb8ed05e4e0d1fe..7b5bb080d278ba5fffbe678841037b71b02b3069 100644 --- a/PaddleRec/multiview_simnet/infer.py +++ b/PaddleRec/multiview_simnet/infer.py @@ -33,32 +33,51 @@ logger.setLevel(logging.INFO) def parse_args(): parser = argparse.ArgumentParser("multi-view simnet") - parser.add_argument( - "--train_file", type=str, help="Training file") - parser.add_argument( - "--valid_file", type=str, help="Validation file") + parser.add_argument("--train_file", type=str, help="Training file") + parser.add_argument("--valid_file", type=str, help="Validation file") parser.add_argument( "--epochs", type=int, default=10, help="Number of epochs for training") parser.add_argument( - "--model_dir", type=str, default='model_output', help="Model output folder") + "--model_dir", + type=str, + default='model_output', + help="Model output folder") parser.add_argument( "--query_slots", type=int, default=1, help="Number of query slots") parser.add_argument( "--title_slots", type=int, default=1, help="Number of title slots") parser.add_argument( - "--query_encoder", type=str, default="bow", help="Encoder module for slot encoding") + "--query_encoder", + type=str, + default="bow", + help="Encoder module for slot encoding") parser.add_argument( - "--title_encoder", type=str, default="bow", help="Encoder module for slot encoding") + "--title_encoder", + type=str, + default="bow", + help="Encoder module for slot encoding") parser.add_argument( - "--query_encode_dim", type=int, default=128, help="Dimension of query encoder output") + "--query_encode_dim", + type=int, + default=128, + help="Dimension of query encoder output") parser.add_argument( - "--title_encode_dim", type=int, default=128, help="Dimension of title encoder output") + "--title_encode_dim", + type=int, + default=128, + help="Dimension of title encoder output") parser.add_argument( "--batch_size", type=int, default=128, help="Batch size for training") parser.add_argument( - "--embedding_dim", type=int, default=128, help="Default Dimension of Embedding") + "--embedding_dim", + type=int, + default=128, + help="Default Dimension of Embedding") parser.add_argument( - "--sparse_feature_dim", type=int, default=1000001, help="Sparse feature hashing space for index processing") + "--sparse_feature_dim", + type=int, + default=1000001, + help="Sparse feature hashing space for index processing") parser.add_argument( "--hidden_size", type=int, default=128, help="Hidden dim") return parser.parse_args() @@ -73,13 +92,14 @@ def start_infer(args, model_path): batch_size=args.batch_size) place = fluid.CPUPlace() exe = fluid.Executor(place) - - with fluid.scope_guard(fluid.core.Scope()): + + with fluid.scope_guard(fluid.Scope()): infer_program, feed_target_names, fetch_vars = fluid.io.load_inference_model( - args.model_dir, exe) + args.model_dir, exe) t0 = time.time() step_id = 0 - feeder = fluid.DataFeeder(program=infer_program, feed_list=feed_target_names, place=place) + feeder = fluid.DataFeeder( + program=infer_program, feed_list=feed_target_names, place=place) for batch_id, data in enumerate(test_reader()): step_id += 1 loss_val, correct_val = exe.run(infer_program, @@ -89,9 +109,11 @@ def start_infer(args, model_path): .format(step_id, batch_id, loss_val, float(correct_val) / args.batch_size)) + def main(): args = parse_args() start_infer(args, args.model_dir) + if __name__ == "__main__": main() diff --git a/PaddleRec/ssr/infer.py b/PaddleRec/ssr/infer.py index 38fb5cd762e117409b12ce8bd202f110a1cdfcb4..4dec0c0d0a09709c76da237445372f3d982e8f8c 100644 --- a/PaddleRec/ssr/infer.py +++ b/PaddleRec/ssr/infer.py @@ -76,7 +76,7 @@ def infer(args, vocab_size, test_reader): hid_size = args.hid_size batch_size = args.batch_size model_path = args.model_dir - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): main_program = fluid.Program() start_up_program = fluid.Program() with fluid.program_guard(main_program, start_up_program): diff --git a/PaddleRec/tagspace/infer.py b/PaddleRec/tagspace/infer.py index 430aa769dab50f8432fe847de01904f545faf1e5..e8522b095826622721de9f2e329c8c361f6f7c41 100644 --- a/PaddleRec/tagspace/infer.py +++ b/PaddleRec/tagspace/infer.py @@ -39,7 +39,7 @@ def infer(test_reader, vocab_tag, use_cuda, model_path, epoch): place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace() exe = fluid.Executor(place) - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): infer_program, feed_target_names, fetch_vars = fluid.io.load_inference_model( model_path, exe) t0 = time.time() diff --git a/PaddleRec/word2vec/infer.py b/PaddleRec/word2vec/infer.py index 9a3649506c546b6ab91cf6f96ff063700679c43d..26a1f2215e96d3c8ee325040f4c76ff93b8a6656 100644 --- a/PaddleRec/word2vec/infer.py +++ b/PaddleRec/word2vec/infer.py @@ -60,7 +60,7 @@ def infer_epoch(args, vocab_size, test_reader, use_cuda, i2w): exe = fluid.Executor(place) emb_size = args.emb_size batch_size = args.batch_size - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): main_program = fluid.Program() with fluid.program_guard(main_program): values, pred = net.infer_network(vocab_size, emb_size) @@ -123,7 +123,7 @@ def infer_step(args, vocab_size, test_reader, use_cuda, i2w): exe = fluid.Executor(place) emb_size = args.emb_size batch_size = args.batch_size - with fluid.scope_guard(fluid.core.Scope()): + with fluid.scope_guard(fluid.Scope()): main_program = fluid.Program() with fluid.program_guard(main_program): values, pred = net.infer_network(vocab_size, emb_size)