From 9a3fb2beac8734a57ad80f926762dfd603391e31 Mon Sep 17 00:00:00 2001 From: xiaoting <31891223+tink2123@users.noreply.github.com> Date: Mon, 17 Feb 2020 12:29:40 +0800 Subject: [PATCH] remove num_workers (#4295) --- dygraph/yolov3/reader.py | 3 +-- dygraph/yolov3/train.py | 1 - dygraph/yolov3/utility.py | 1 - 3 files changed, 1 insertion(+), 4 deletions(-) diff --git a/dygraph/yolov3/reader.py b/dygraph/yolov3/reader.py index c74a938e..1c0d76a3 100644 --- a/dygraph/yolov3/reader.py +++ b/dygraph/yolov3/reader.py @@ -296,8 +296,7 @@ def train(size=416, total_iter=0, mixup_iter=0, random_sizes=[], - num_workers=8, - max_queue=32, + num_workers=1, use_multiprocess_reader=True, use_gpu=True): generator = dsr.get_reader('train', size, batch_size, shuffle, shuffle_seed, diff --git a/dygraph/yolov3/train.py b/dygraph/yolov3/train.py index 7c742fd3..b09455a5 100755 --- a/dygraph/yolov3/train.py +++ b/dygraph/yolov3/train.py @@ -142,7 +142,6 @@ def train(): mixup_iter=mixup_iter * devices_num, random_sizes=random_sizes, use_multiprocess_reader=cfg.use_multiprocess_reader, - num_workers=cfg.worker_num, use_gpu=cfg.use_gpu) if args.use_data_parallel: train_reader = fluid.contrib.reader.distributed_batch_reader(train_reader) diff --git a/dygraph/yolov3/utility.py b/dygraph/yolov3/utility.py index 3b0b4a59..fa9b0a94 100644 --- a/dygraph/yolov3/utility.py +++ b/dygraph/yolov3/utility.py @@ -123,7 +123,6 @@ def parse_args(): add_arg('data_dir', str, 'dataset/coco', "The data root path.") add_arg('start_iter', int, 0, "Start iteration.") add_arg('use_multiprocess_reader', bool, True, "whether use multiprocess reader.") - add_arg('worker_num', int, 8, "worker number for multiprocess reader.") add_arg('use_data_parallel', ast.literal_eval, False, "the flag indicating whether to use data parallel model to train the model") #SOLVER add_arg('batch_size', int, 8, "Mini-batch size per device.") -- GitLab