diff --git a/paddle/fluid/operators/optimizers/adam_op.h b/paddle/fluid/operators/optimizers/adam_op.h index 7dd5a8783a5479d0fba2794f3121cfbd5041b016..5ba5639fd51a639e878e650585b8e7e29317b3c1 100644 --- a/paddle/fluid/operators/optimizers/adam_op.h +++ b/paddle/fluid/operators/optimizers/adam_op.h @@ -358,7 +358,7 @@ class AdamOpKernel : public framework::OpKernel { FLAGS_min_param_size_to_use_multithread > 0 && param.numel() > FLAGS_min_param_size_to_use_multithread) { VLOG(3) << "use multi thread, inner_op_parallelism=" - << inner_op_parallelism << " min_param_size_to_use_multithread" + << inner_op_parallelism << " min_param_size_to_use_multithread=" << FLAGS_min_param_size_to_use_multithread; std::vector> fs; int64_t block_size = param.numel() / inner_op_parallelism;