提交 0d4693bd 编写于 作者: Q Qiao Longfei

use logging

上级 a08c0691
import argparse import argparse
import time import logging
import numpy as np import numpy as np
import paddle import paddle
...@@ -9,9 +9,10 @@ import reader ...@@ -9,9 +9,10 @@ import reader
from network_conf import ctr_dnn_model from network_conf import ctr_dnn_model
def print_log(log_str): logging.basicConfig(
time_stamp = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(time.time())) format='%(asctime)s - %(levelname)s - %(message)s')
print(str(time_stamp) + " " + log_str) logger = logging.getLogger("fluid")
logger.setLevel(logging.INFO)
def parse_args(): def parse_args():
...@@ -70,7 +71,7 @@ def infer(): ...@@ -70,7 +71,7 @@ def infer():
feed=feeder.feed(data), feed=feeder.feed(data),
fetch_list=fetch_targets) fetch_list=fetch_targets)
if batch_id % 100 == 0: if batch_id % 100 == 0:
print_log("TEST --> batch: {} loss: {} auc: {}".format(batch_id, loss_val, auc_val)) logger.info("TEST --> batch: {} loss: {} auc: {}".format(batch_id, loss_val, auc_val))
if __name__ == '__main__': if __name__ == '__main__':
......
from __future__ import print_function from __future__ import print_function
import argparse import argparse
import logging
import os import os
import time
import paddle import paddle
import paddle.fluid as fluid import paddle.fluid as fluid
...@@ -10,10 +10,10 @@ import paddle.fluid as fluid ...@@ -10,10 +10,10 @@ import paddle.fluid as fluid
import reader import reader
from network_conf import ctr_dnn_model from network_conf import ctr_dnn_model
logging.basicConfig(
def print_log(log_str): format='%(asctime)s - %(levelname)s - %(message)s')
time_stamp = time.strftime('%Y-%m-%d %H:%M:%S', time.localtime(time.time())) logger = logging.getLogger("fluid")
print(str(time_stamp) + " " + log_str) logger.setLevel(logging.INFO)
def parse_args(): def parse_args():
...@@ -105,7 +105,7 @@ def train_loop(args, train_program, data_list, loss, auc_var, batch_auc_var): ...@@ -105,7 +105,7 @@ def train_loop(args, train_program, data_list, loss, auc_var, batch_auc_var):
feed=feeder.feed(data), feed=feeder.feed(data),
fetch_list=[loss, auc_var, batch_auc_var] fetch_list=[loss, auc_var, batch_auc_var]
) )
print_log("TRAIN --> pass: {} batch: {} loss: {} auc: {}, batch_auc: {}" logger.info("TRAIN --> pass: {} batch: {} loss: {} auc: {}, batch_auc: {}"
.format(pass_id, batch_id, loss_val/args.batch_size, auc_val, batch_auc_val)) .format(pass_id, batch_id, loss_val/args.batch_size, auc_val, batch_auc_val))
if batch_id % 1000 == 0 and batch_id != 0: if batch_id % 1000 == 0 and batch_id != 0:
model_dir = args.model_output_dir + '/batch-' + str(batch_id) model_dir = args.model_output_dir + '/batch-' + str(batch_id)
...@@ -127,22 +127,22 @@ def train(): ...@@ -127,22 +127,22 @@ def train():
optimizer.minimize(loss) optimizer.minimize(loss)
if args.is_local: if args.is_local:
print_log("run local training") logger.info("run local training")
main_program = fluid.default_main_program() main_program = fluid.default_main_program()
train_loop(args, main_program, data_list, loss, auc_var, batch_auc_var) train_loop(args, main_program, data_list, loss, auc_var, batch_auc_var)
else: else:
print_log("run dist training") logger.info("run dist training")
t = fluid.DistributeTranspiler() t = fluid.DistributeTranspiler()
t.transpile(args.trainer_id, pservers=args.endpoints, trainers=args.trainers) t.transpile(args.trainer_id, pservers=args.endpoints, trainers=args.trainers)
if args.role == "pserver": if args.role == "pserver":
print_log("run pserver") logger.info("run pserver")
prog = t.get_pserver_program(args.current_endpoint) prog = t.get_pserver_program(args.current_endpoint)
startup = t.get_startup_program(args.current_endpoint, pserver_program=prog) startup = t.get_startup_program(args.current_endpoint, pserver_program=prog)
exe = fluid.Executor(fluid.CPUPlace()) exe = fluid.Executor(fluid.CPUPlace())
exe.run(startup) exe.run(startup)
exe.run(prog) exe.run(prog)
elif args.role == "trainer": elif args.role == "trainer":
print_log("run trainer") logger.info("run trainer")
train_prog = t.get_trainer_program() train_prog = t.get_trainer_program()
train_loop(args, train_prog, data_list, loss, auc_var, batch_auc_var) train_loop(args, train_prog, data_list, loss, auc_var, batch_auc_var)
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册