From 88ae3f169db7403ba99685b0293dece4c4b6b091 Mon Sep 17 00:00:00 2001 From: Xin Pan Date: Thu, 20 Sep 2018 09:25:35 +0800 Subject: [PATCH] further clean --- benchmark/fluid/args.py | 4 ---- benchmark/fluid/models/resnet.py | 5 ----- benchmark/fluid/models/resnet_with_preprocess.py | 5 ----- benchmark/fluid/models/se_resnext.py | 8 +------- python/paddle/fluid/optimizer.py | 1 - python/paddle/fluid/tests/book/test_recognize_digits.py | 2 +- 6 files changed, 2 insertions(+), 23 deletions(-) diff --git a/benchmark/fluid/args.py b/benchmark/fluid/args.py index 0d5c9652de6..9540900b112 100644 --- a/benchmark/fluid/args.py +++ b/benchmark/fluid/args.py @@ -136,10 +136,6 @@ def parse_args(): '--no_random', action='store_true', help='If set, keep the random seed and do not shuffle the data.') - parser.add_argument( - '--use_lars', - action='store_true', - help='If set, use lars for optimizers, ONLY support resnet module.') parser.add_argument( '--reduce_strategy', type=str, diff --git a/benchmark/fluid/models/resnet.py b/benchmark/fluid/models/resnet.py index 1b3bfe659c7..f692e7722a1 100644 --- a/benchmark/fluid/models/resnet.py +++ b/benchmark/fluid/models/resnet.py @@ -200,11 +200,6 @@ def get_model(args, is_train, main_prog, startup_prog): # configure optimize optimizer = None if is_train: - if args.use_lars: - lars_decay = 1.0 - else: - lars_decay = 0.0 - total_images = 1281167 / trainer_count step = int(total_images / (args.batch_size * args.gpus) + 1) diff --git a/benchmark/fluid/models/resnet_with_preprocess.py b/benchmark/fluid/models/resnet_with_preprocess.py index e8d661d8475..e996c9a7045 100644 --- a/benchmark/fluid/models/resnet_with_preprocess.py +++ b/benchmark/fluid/models/resnet_with_preprocess.py @@ -224,11 +224,6 @@ def get_model(args, is_train, main_prog, startup_prog): # configure optimize optimizer = None if is_train: - if args.use_lars: - lars_decay = 1.0 - else: - lars_decay = 0.0 - total_images = 1281167 / trainer_count step = int(total_images / args.batch_size + 1) diff --git a/benchmark/fluid/models/se_resnext.py b/benchmark/fluid/models/se_resnext.py index 9f887fb324d..7fbb83c2ec1 100644 --- a/benchmark/fluid/models/se_resnext.py +++ b/benchmark/fluid/models/se_resnext.py @@ -244,11 +244,6 @@ def get_model(args, is_train, main_prog, startup_prog): optimizer = None if is_train: - if args.use_lars: - lars_decay = 1.0 - else: - lars_decay = 0.0 - total_images = 1281167 / trainer_count step = int(total_images / args.batch_size + 1) @@ -262,8 +257,7 @@ def get_model(args, is_train, main_prog, startup_prog): learning_rate=fluid.layers.piecewise_decay( boundaries=bd, values=lr), momentum=0.9, - regularization=fluid.regularizer.L2Decay(1e-4), - LARS_weight_decay=lars_decay) + regularization=fluid.regularizer.L2Decay(1e-4)) optimizer.minimize(avg_cost) if args.memory_optimize: diff --git a/python/paddle/fluid/optimizer.py b/python/paddle/fluid/optimizer.py index 3c8c83560de..ad09005d866 100644 --- a/python/paddle/fluid/optimizer.py +++ b/python/paddle/fluid/optimizer.py @@ -104,7 +104,6 @@ class Optimizer(object): param = param_and_grad[0] param_lr = param.optimize_attr['learning_rate'] if type(param_lr) == Variable: - # param learning rate has been updated (LARS) print("returns updated param lr ", param_lr) return param_lr else: diff --git a/python/paddle/fluid/tests/book/test_recognize_digits.py b/python/paddle/fluid/tests/book/test_recognize_digits.py index 135f11d24c8..4b4f3e40377 100644 --- a/python/paddle/fluid/tests/book/test_recognize_digits.py +++ b/python/paddle/fluid/tests/book/test_recognize_digits.py @@ -99,7 +99,7 @@ def train(nn_type, test_program = fluid.default_main_program().clone(for_test=True) - optimizer = fluid.optimizer.Adam(learning_rate=0.001, LARS_weight_decay=0.3) + optimizer = fluid.optimizer.Adam(learning_rate=0.001) optimizer.minimize(avg_loss) place = fluid.CUDAPlace(0) if use_cuda else fluid.CPUPlace() -- GitLab