From 4d4593b91392a2a1414a2cceca7cea62879d01ee Mon Sep 17 00:00:00 2001 From: Yibing Liu Date: Mon, 22 May 2017 18:24:54 +0800 Subject: [PATCH] code cleanup --- paddle/parameter/FirstOrderOptimizer.cpp | 3 ++- paddle/parameter/ParameterOptimizer.h | 8 -------- 2 files changed, 2 insertions(+), 9 deletions(-) diff --git a/paddle/parameter/FirstOrderOptimizer.cpp b/paddle/parameter/FirstOrderOptimizer.cpp index 02e600adb9e..207fb33f4e8 100644 --- a/paddle/parameter/FirstOrderOptimizer.cpp +++ b/paddle/parameter/FirstOrderOptimizer.cpp @@ -305,12 +305,13 @@ void AdamaxParameterOptimizer::update(const VectorPtr vecs[], void OptimizerWithGradientClipping::update(const VectorPtr vecs[], const ParameterConfig& config, size_t sparseId) const { - // globalGradientClipping(vecs, config, FLAGS_log_clipping); real global_thres_ = optConfig_.gradient_clipping_threshold(); real local_thres_ = config.gradient_clipping_threshold(); real threshold; std::string field; + // Get the minimum of local and global threshold + // as the real threshold for clipping if (global_thres_ > 0.0f && local_thres_ > 0.0f) { threshold = global_thres_ < local_thres_ ? global_thres_ : local_thres_; field = global_thres_ < local_thres_ ? "global" : "local"; diff --git a/paddle/parameter/ParameterOptimizer.h b/paddle/parameter/ParameterOptimizer.h index 38d432ba9bc..f98ba569b56 100644 --- a/paddle/parameter/ParameterOptimizer.h +++ b/paddle/parameter/ParameterOptimizer.h @@ -170,9 +170,6 @@ public: real getLearningRate() const { return learningRate_; } - // real getGradientClippingThreshold() const {return - // gradientClippingThreshold_;} - virtual void setNoDecay() { applyDecay_ = false; } static ParameterOptimizer* create(const OptimizationConfig& optConfig, @@ -206,11 +203,6 @@ protected: */ real learningRate_; - /** - * global threshold for gradient clipping, - * init value is opt_config.gradient_clipping_thresholod - */ - std::unique_ptr learningRateScheduler_; int64_t pass_; // current training pass (starting from 0) bool firstTime_; -- GitLab