diff --git a/pdseg/eval.py b/pdseg/eval.py index c34a8b5e82aa7dc5a67d5b7386ca833917e334d8..815ebf56f09140a5911bdeb1aa4acd3dbf7950d1 100644 --- a/pdseg/eval.py +++ b/pdseg/eval.py @@ -102,7 +102,7 @@ def evaluate(cfg, ckpt_dir=None, use_gpu=False, use_mpio=False, **kwargs): places = fluid.cuda_places() if use_gpu else fluid.cpu_places() place = places[0] dev_count = len(places) - print("Device count = {}".format(dev_count)) + print("#Device count: {}".format(dev_count)) exe = fluid.Executor(place) exe.run(startup_prog) diff --git a/pdseg/train.py b/pdseg/train.py index 05db0de833d76e43643a46a4b6da328b29f05f82..8738e43ccd95d631efa93e06111c8292ae6d6303 100644 --- a/pdseg/train.py +++ b/pdseg/train.py @@ -216,7 +216,7 @@ def train(cfg): place = places[0] # Get number of GPU dev_count = len(places) - print("#GPU-Devices: {}".format(dev_count)) + print("#Device count: {}".format(dev_count)) # Make sure BATCH_SIZE can divided by GPU cards assert cfg.BATCH_SIZE % dev_count == 0, (