diff --git a/dygraph/yolov3/reader.py b/dygraph/yolov3/reader.py index c74a938e1b7f8a40737e60b1547cec222719ee15..1c0d76a3454edc7b16281e4209c46b3df2c894e3 100644 --- a/dygraph/yolov3/reader.py +++ b/dygraph/yolov3/reader.py @@ -296,8 +296,7 @@ def train(size=416, total_iter=0, mixup_iter=0, random_sizes=[], - num_workers=8, - max_queue=32, + num_workers=1, use_multiprocess_reader=True, use_gpu=True): generator = dsr.get_reader('train', size, batch_size, shuffle, shuffle_seed, diff --git a/dygraph/yolov3/train.py b/dygraph/yolov3/train.py index 7c742fd30006c4d7a99d44611d4f18b4f89949bf..b09455a5babc1339ebae7654122ccbe027cd7abb 100755 --- a/dygraph/yolov3/train.py +++ b/dygraph/yolov3/train.py @@ -142,7 +142,6 @@ def train(): mixup_iter=mixup_iter * devices_num, random_sizes=random_sizes, use_multiprocess_reader=cfg.use_multiprocess_reader, - num_workers=cfg.worker_num, use_gpu=cfg.use_gpu) if args.use_data_parallel: train_reader = fluid.contrib.reader.distributed_batch_reader(train_reader) diff --git a/dygraph/yolov3/utility.py b/dygraph/yolov3/utility.py index 3b0b4a59efc99e020148c9049d58e447f271ed49..fa9b0a945f0747b0f9243ca7422a45413702b92a 100644 --- a/dygraph/yolov3/utility.py +++ b/dygraph/yolov3/utility.py @@ -123,7 +123,6 @@ def parse_args(): add_arg('data_dir', str, 'dataset/coco', "The data root path.") add_arg('start_iter', int, 0, "Start iteration.") add_arg('use_multiprocess_reader', bool, True, "whether use multiprocess reader.") - add_arg('worker_num', int, 8, "worker number for multiprocess reader.") add_arg('use_data_parallel', ast.literal_eval, False, "the flag indicating whether to use data parallel model to train the model") #SOLVER add_arg('batch_size', int, 8, "Mini-batch size per device.")