backward.cc 17.6 KB
Newer Older
1
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserved.
Y
Yu Yang 已提交
2

L
Luo Tao 已提交
3 4 5
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
Y
Yu Yang 已提交
6

L
Luo Tao 已提交
7
    http://www.apache.org/licenses/LICENSE-2.0
Y
Yu Yang 已提交
8

L
Luo Tao 已提交
9 10 11 12 13
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
Y
Yu Yang 已提交
14

Y
Yi Wang 已提交
15
#include "paddle/fluid/framework/backward.h"
D
dongzhihong 已提交
16

F
fengjiayi 已提交
17
#include <deque>
D
dongzhihong 已提交
18
#include <list>
Y
Yu Yang 已提交
19
#include <memory>
Y
Yu Yang 已提交
20
#include <unordered_set>
Y
Yu Yang 已提交
21

Y
Yi Wang 已提交
22 23
#include "paddle/fluid/framework/block_desc.h"
#include "paddle/fluid/framework/op_registry.h"
Y
Yu Yang 已提交
24 25 26 27

namespace paddle {
namespace framework {

28 29 30 31 32 33
static std::unordered_set<std::string>* g_ctrl_flow_ops_ = nullptr;
// Control Flow operators's backward is significantly different from
// computational operators. Hack Code here.
// We should design a better way to backward CtrlFlowOps.
static std::unordered_set<std::string>& CtrlFlowOps() {
  if (g_ctrl_flow_ops_ == nullptr) {
34 35
    g_ctrl_flow_ops_ = new std::unordered_set<std::string>{
        "increment", "lod_rank_table", "less_than"};
36 37 38 39
  }
  return *g_ctrl_flow_ops_;
}

Y
Yu Yang 已提交
40
static inline std::unique_ptr<OperatorBase> CreateGradOp(
41 42
    const OperatorBase& op, const std::unordered_set<std::string>& no_grad_set,
    std::unordered_map<std::string, std::string>* grad_to_var) {
Y
Yu Yang 已提交
43
  OpDesc op_desc;
Y
Yu Yang 已提交
44 45 46 47 48
  op_desc.SetInputMap(op.Inputs());
  op_desc.SetOutputMap(op.Outputs());
  op_desc.SetType(op.Type());
  op_desc.SetAttrMap(op.Attrs());
  auto& info = OpInfoMap::Instance().Get(op.Type());
Y
Yu Yang 已提交
49
  auto grad_descs = info.GradOpMaker()(op_desc, no_grad_set, grad_to_var, {});
Y
Yu Yang 已提交
50 51
  std::vector<std::unique_ptr<OperatorBase>> grad_ops;
  grad_ops.reserve(grad_descs.size());
Y
Yu Yang 已提交
52 53
  std::transform(grad_descs.begin(), grad_descs.end(),
                 std::back_inserter(grad_ops),
Y
Yu Yang 已提交
54
                 [](const std::unique_ptr<OpDesc>& grad_desc) {
Y
Yu Yang 已提交
55
                   return OpRegistry::CreateOp(*grad_desc);
Y
Yu Yang 已提交
56
                 });
Y
Yu Yang 已提交
57
  PADDLE_ENFORCE(!grad_ops.empty());
Y
Yu Yang 已提交
58 59 60
  if (grad_ops.size() == 1) {
    return std::move(grad_ops[0]);
  } else {
61
    PADDLE_THROW("Unexpected Branch");
Y
Yu Yang 已提交
62 63 64
  }
}

Y
Yu Yang 已提交
65
template <typename Map, typename T>
Q
qiaolongfei 已提交
66
static void ForEachVarName(const Map& names, T callback) {
Y
Yu Yang 已提交
67
  for (auto& name : names) {
Y
Yu Yang 已提交
68
    for (auto& n : name.second) {
69
      if (callback(n)) return;
Y
Yu Yang 已提交
70 71
    }
  }
Y
Yu Yang 已提交
72 73
}

Y
Yan Chunwei 已提交
74
// return whether all the names + suffixes in the set
Y
Yu Yang 已提交
75
static bool AllInSet(
Y
Yu Yang 已提交
76
    const std::map<std::string, std::vector<std::string>>& names,
Y
Yu Yang 已提交
77
    const std::string& suffix, const std::unordered_set<std::string>& set) {
78 79 80 81
  bool all_in_set = true;
  ForEachVarName(names, [&all_in_set, &set, &suffix](const std::string& n) {
    all_in_set = set.find(n + suffix) != set.end();
    return !all_in_set;
Y
Yu Yang 已提交
82
  });
83
  return all_in_set;
Y
Yu Yang 已提交
84 85
}

Y
Yu Yang 已提交
86
static std::unique_ptr<OperatorBase> NOP() {
87
  PADDLE_THROW("Unexpected Branch");
Y
Yu Yang 已提交
88 89
}

Y
Yan Chunwei 已提交
90
//  Get backward operator from a forward operator, a recursive implementation.
Y
Yu Yang 已提交
91 92 93
//
//  no_grad_names the gradient variable names without gradient calculating.
//
94 95 96
//  uniq_id is a unique index used inside recursively calling
//  BackwardRecursive. use `uid = uniq_id++;` to get the unique index, and
//  pass `uniq_id` through recursive calling.
Y
Yu Yang 已提交
97
//
Y
Yan Chunwei 已提交
98 99
//  returns The backward operator. In a simple situation, it may be a simple
//  operator, in a complex situation, it maybe a NetOp.
Y
Yu Yang 已提交
100 101
//
//  See Backward.h for details
Y
Yu Yang 已提交
102
static std::unique_ptr<OperatorBase> BackwardRecursive(
Y
Yu Yang 已提交
103
    const OperatorBase& forwardOp,
104 105 106
    std::unordered_set<std::string>& no_grad_names,
    std::unordered_map<std::string, std::string>* grad_to_var,
    size_t& uniq_id) {
Y
Yu Yang 已提交
107 108
  //  If all input gradients of forwarding operator do not need to calculate,
  //  just return an NOP. Not return null ptr because NOP does not take
Q
typo  
qiaolongfei 已提交
109
  //  too much time for calculation, but it is useful for simplifying logic.
110
  if (AllInSet(forwardOp.Inputs() /*names*/, kGradVarSuffix /*suffix*/,
Y
Yan Chunwei 已提交
111
               no_grad_names /*set*/)) {
Y
Yu Yang 已提交
112
    return NOP();
Y
Yu Yang 已提交
113 114
  }

115 116
  //  All output gradients of forwarding operator do not need to calculate.
  //  Then all input gradients cannot be computed at all, and we put them into
Y
Yu Yang 已提交
117
  //  `no_grad_names` set. Return an NOP.
Q
qiaolongfei 已提交
118
  if (AllInSet(forwardOp.Outputs() /*names*/, kGradVarSuffix /*suffix*/,
Y
Yan Chunwei 已提交
119
               no_grad_names /*set*/)) {
Q
qiaolongfei 已提交
120
    ForEachVarName(forwardOp.Inputs(),
Y
Yu Yang 已提交
121 122 123 124
                   [&no_grad_names](const std::string& name) -> bool {
                     no_grad_names.insert(GradVarName(name));
                     return false;
                   });
Y
Yu Yang 已提交
125
    return NOP();
Y
Yu Yang 已提交
126 127
  }

Y
Yu Yang 已提交
128
  // Returned gradient network
129
  PADDLE_THROW("Unexpected Branch");
Y
Yu Yang 已提交
130
}
Y
Yu Yang 已提交
131

Y
Yu Yang 已提交
132
// See header for comments
Y
Yu Yang 已提交
133
std::unique_ptr<OperatorBase> Backward(
Y
Yu Yang 已提交
134
    const OperatorBase& forwardOp,
Y
Yu Yang 已提交
135 136
    const std::unordered_set<std::string>& no_grad_vars) {
  std::unordered_set<std::string> no_grad_names;
Q
qijun 已提交
137
  no_grad_names.reserve(no_grad_vars.size() + 1);
Y
Yu Yang 已提交
138

139
  no_grad_names.insert(std::string(kEmptyVarName) + kGradVarSuffix);
140

Y
Yu Yang 已提交
141
  for (auto& name : no_grad_vars) {
142
    no_grad_names.insert(name + kGradVarSuffix);
Y
Yu Yang 已提交
143
  }
Y
Yu Yang 已提交
144
  size_t uid = 0;
145 146
  std::unordered_map<std::string, std::string> grad_to_var;
  return BackwardRecursive(forwardOp, no_grad_names, &grad_to_var, uid);
Y
Yu Yang 已提交
147
}
Y
Yi Wang 已提交
148

F
fengjiayi 已提交
149 150 151 152 153 154 155 156 157
// ====================================  //

static bool AllGradInSet(const std::vector<std::string>& names,
                         const std::unordered_set<std::string>& set) {
  for (const std::string& name : names) {
    if (!set.count(GradVarName(name))) {
      return false;
    }
  }
Y
Yang Yang(Tony) 已提交
158 159 160 161 162 163 164 165 166 167 168 169 170
  if (VLOG_IS_ON(10)) {
    std::ostringstream sout;
    sout << "All input {";
    for (auto& name : names) {
      sout << name << ",";
    }
    sout << "} is in {";
    for (auto& name : set) {
      sout << name << ",";
    }
    sout << "}";
    VLOG(10) << sout.str();
  }
F
fengjiayi 已提交
171 172 173
  return true;
}

Y
Yu Yang 已提交
174 175 176 177 178 179 180 181 182
static std::string FwdName(const std::string& grad_name) {
  auto pos = grad_name.find("@GRAD");
  if (pos == std::string::npos) {
    return "";
  } else {
    return grad_name.substr(0, pos);
  }
}

Y
Yu Yang 已提交
183
static void CreateGradVarInBlock(
184 185
    size_t grad_op_start_index,
    const std::unordered_map<std::string, std::string>& param_name_map,
Y
Yu Yang 已提交
186
    BlockDesc* block_desc,
187
    std::unordered_map<std::string, GradVarInfo>* grad_var_record) {
188 189 190
  auto ops = block_desc->AllOps();
  for (size_t op_index = grad_op_start_index; op_index < ops.size();
       ++op_index) {
Y
Yu Yang 已提交
191
    std::unordered_set<std::string> new_vars;
192
    auto& ctrl_flow_ops = CtrlFlowOps();
Y
Yu Yang 已提交
193 194
    ForEachVarName(ops[op_index]->Outputs(),
                   [&](const std::string& grad_var_name) {
195 196 197 198 199 200 201 202 203 204 205
                     if (ctrl_flow_ops.find(ops[op_index]->Type()) !=
                         ctrl_flow_ops.end()) {
                       if (block_desc->HasVarRecursive(grad_var_name)) {
                         return false;
                       }
                     } else {
                       if (block_desc->HasVar(grad_var_name)) {
                         return false;
                       }
                     }
                     if (grad_var_name == framework::kEmptyVarName) {
Y
Yu Yang 已提交
206 207
                       return false;
                     }
Q
Qiao Longfei 已提交
208
                     auto var = block_desc->Var(grad_var_name);
209
                     VLOG(10) << "Creating Variable " << grad_var_name;
Y
Yu Yang 已提交
210
                     new_vars.insert(var->Name());
Y
Yu Yang 已提交
211 212 213 214 215 216 217 218 219 220 221
                     auto it = param_name_map.find(grad_var_name);
                     if (it == param_name_map.end()) {
                       return false;
                     }
                     auto param_var_name = it->second;
                     auto& grad_record = (*grad_var_record)[param_var_name];
                     grad_record.name_ = grad_var_name;
                     grad_record.block_idx_ = block_desc->ID();
                     grad_record.op_idx_ = static_cast<int>(op_index);
                     return false; /* not break */
                   });
Y
Yang Yang(Tony) 已提交
222 223 224 225 226 227 228 229 230
    ops[op_index]->InferVarType(block_desc);
    for (auto& arg : ops[op_index]->OutputArgumentNames()) {
      if (new_vars.find(arg) == new_vars.end()) {
        continue;
      }
      auto pname = FwdName(arg);
      auto* param = block_desc->FindVarRecursive(pname);
      auto* grad = block_desc->FindVar(arg);
      if (param == nullptr) {
231
        grad->SetDataType(proto::VarType::FP32);
Y
Yang Yang(Tony) 已提交
232 233
      } else {
        grad->SetDataType(param->GetDataType());
Y
Yu Yang 已提交
234
      }
Q
Qiao Longfei 已提交
235
    }
Y
Yang Yang(Tony) 已提交
236
    ops[op_index]->InferShape(*block_desc);
237 238 239
  }
}

Y
Yu Yang 已提交
240 241
std::vector<std::unique_ptr<OpDesc>> MakeOpGrad(
    const OpDesc* op_desc, std::unordered_set<std::string>* no_grad_vars,
Y
Yu Yang 已提交
242
    std::unordered_map<std::string, std::string>* grad_to_var,
Y
Yu Yang 已提交
243 244
    const std::vector<BlockDesc*>& grad_block = std::vector<BlockDesc*>()) {
  std::vector<std::unique_ptr<OpDesc>> grad_op_descs;
245
  // All input gradients of forwarding operator do not need to calculate.
F
fengjiayi 已提交
246
  const std::vector<std::string>& inputs = op_desc->InputArgumentNames();
247
  if (AllGradInSet(inputs, *no_grad_vars)) {
248
    VLOG(10) << "Drop operator  " << op_desc->Type();
F
fengjiayi 已提交
249 250
    return grad_op_descs;  // empty vector
  }
251

F
fengjiayi 已提交
252
  // All output gradients of forwarding operator do not need to calculate.
F
fengjiayi 已提交
253
  const std::vector<std::string>& outputs = op_desc->OutputArgumentNames();
254

255
  if (AllGradInSet(outputs, *no_grad_vars)) {
256 257 258 259 260 261 262 263 264
    VLOG(10) << "Drop operator " << op_desc->Type();
    // FIXME: Hack code here
    auto& ctrl_flow_ops = CtrlFlowOps();
    if (ctrl_flow_ops.find(op_desc->Type()) == ctrl_flow_ops.end()) {
      // Only computational op need drop input's gradient.
      for (const std::string& name : inputs) {
        no_grad_vars->insert(GradVarName(name));
        VLOG(10) << " Also drop " << GradVarName(name);
      }
F
fengjiayi 已提交
265
    }
266

F
fengjiayi 已提交
267 268 269
    return grad_op_descs;  // empty vector
  }

Y
Yu Yang 已提交
270 271 272 273
  grad_op_descs =
      OpInfoMap::Instance()
          .Get(op_desc->Type())
          .GradOpMaker()(*op_desc, *no_grad_vars, grad_to_var, grad_block);
F
fengjiayi 已提交
274

Y
Yu Yang 已提交
275
  std::list<std::unique_ptr<OpDesc>> pending_fill_zeros_ops;
F
Update  
fengjiayi 已提交
276 277
  for (auto& desc : grad_op_descs) {
    for (const std::string& in_name : desc->InputArgumentNames()) {
278
      if (no_grad_vars->count(in_name)) {
F
fengjiayi 已提交
279 280 281
        std::string prefix = in_name.substr(
            0, in_name.size() - sizeof(kGradVarSuffix) / sizeof(char) + 1);
        std::string new_name = prefix + kZeroVarSuffix;
F
Update  
fengjiayi 已提交
282
        desc->Rename(in_name, new_name);
Y
Yu Yang 已提交
283
        std::unique_ptr<OpDesc> fill_zeros_op(
F
fengjiayi 已提交
284 285
            new OpDesc("fill_zeros_like", {{"X", {prefix}}},
                       {{"Out", {new_name}}}, AttributeMap{}));
F
fengjiayi 已提交
286
        pending_fill_zeros_ops.push_back(std::move(fill_zeros_op));
F
fengjiayi 已提交
287 288 289
      }
    }
  }
F
fengjiayi 已提交
290

F
fengjiayi 已提交
291
  for (auto& p : pending_fill_zeros_ops) {
F
fengjiayi 已提交
292
    grad_op_descs.insert(grad_op_descs.begin(), std::move(p));
F
fengjiayi 已提交
293
  }
F
fengjiayi 已提交
294 295 296
  return grad_op_descs;
}

Y
Yu Yang 已提交
297 298
static BlockDesc* CreateStepBlock(
    ProgramDesc& program_desc, std::unordered_set<std::string>* no_grad_vars,
Y
Yu Yang 已提交
299 300 301
    std::unordered_map<std::string, std::string>* grad_to_var,
    int step_block_idx);

Y
Yu Yang 已提交
302 303
std::vector<std::unique_ptr<OpDesc>> MakeBlockBackward(
    ProgramDesc& program_desc, int block_idx,
304 305
    std::unordered_set<std::string>* no_grad_vars,
    std::unordered_map<std::string, std::string>* grad_to_var) {
Y
Yang Yang(Tony) 已提交
306
  VLOG(5) << "MakeBlockBackward";
Y
Yu Yang 已提交
307 308
  BlockDesc* cur_block = program_desc.MutableBlock(block_idx);
  std::vector<OpDesc*> op_descs = cur_block->AllOps();
F
Update  
fengjiayi 已提交
309 310
  std::unordered_map<std::string, std::vector<size_t>> dup_out_ops;
  size_t grad_desc_idx = 0;
Y
Yu Yang 已提交
311
  std::vector<std::unique_ptr<OpDesc>> backward_descs;
312

F
fengjiayi 已提交
313
  for (auto it = op_descs.rbegin(); it != op_descs.rend(); ++it) {
Y
Yang Yang(Tony) 已提交
314
    VLOG(5) << "Making backward " << (*it)->Type() << " op";
Y
Yu Yang 已提交
315
    std::vector<std::unique_ptr<OpDesc>> op_grads;
F
fengjiayi 已提交
316

Y
Yang Yang 已提交
317 318
    if ((*it)->Type() == "recurrent" || (*it)->Type() == "while" ||
        (*it)->Type() == "parallel_do") {
319
      int step_block_idx = (*it)->GetBlockAttr("sub_block");
Y
Yu Yang 已提交
320 321
      BlockDesc* backward_block = CreateStepBlock(program_desc, no_grad_vars,
                                                  grad_to_var, step_block_idx);
Y
Yu Yang 已提交
322 323
      op_grads = MakeOpGrad(*it, no_grad_vars, grad_to_var, {backward_block});
    } else if ((*it)->Type() == "conditional_block") {
Y
Yu Yang 已提交
324
      BlockDesc* backward_block =
Y
Yu Yang 已提交
325
          CreateStepBlock(program_desc, no_grad_vars, grad_to_var,
326
                          (*it)->GetBlockAttr("sub_block"));
Y
Yu Yang 已提交
327 328 329
      op_grads = MakeOpGrad(*it, no_grad_vars, grad_to_var, {backward_block});
    } else {
      op_grads = MakeOpGrad(*it, no_grad_vars, grad_to_var);
F
fengjiayi 已提交
330 331
    }

Y
Yang Yang(Tony) 已提交
332 333 334 335 336 337 338 339 340
    if (VLOG_IS_ON(10)) {
      std::ostringstream sout;
      sout << "Made ";
      for (auto& op_grad : op_grads) {
        sout << op_grad->Type() << " ";
      }
      VLOG(10) << sout.str();
    }

F
Update  
fengjiayi 已提交
341
    for (const auto& desc : op_grads) {
F
fengjiayi 已提交
342
      for (const std::string& out_name : desc->OutputArgumentNames()) {
343 344 345 346 347
        if (out_name.find("@GRAD") == std::string::npos) {
          // Not all outputs of a backward operator is a gradient. Only gradient
          // need to be sum. Skip variables are not gradient.
          continue;
        }
F
Update  
fengjiayi 已提交
348 349 350 351
        dup_out_ops[out_name].emplace_back(grad_desc_idx);
      }
      ++grad_desc_idx;
    }
Y
Yu Yang 已提交
352 353 354
    std::transform(op_grads.begin(), op_grads.end(),
                   std::back_inserter(backward_descs),
                   [](std::unique_ptr<OpDesc>& ptr) { return std::move(ptr); });
F
Update  
fengjiayi 已提交
355
  }
Y
Yang Yang(Tony) 已提交
356 357

  VLOG(5) << "Appending Sums";
F
Update  
fengjiayi 已提交
358
  // Check whether some variables are written more than once
Y
Yu Yang 已提交
359
  std::list<std::pair<size_t, std::unique_ptr<OpDesc>>> pending_sum_ops;
F
Update  
fengjiayi 已提交
360 361 362 363 364
  for (const auto& dup : dup_out_ops) {
    const std::string& out_name = dup.first;
    const std::vector<size_t> dup_op = dup.second;
    if (out_name != kEmptyVarName && dup_op.size() > 1) {
      std::vector<std::string> sum_op_inputs;
Y
Yang Yang(Tony) 已提交
365
      std::string next_g_name = out_name;
F
Update  
fengjiayi 已提交
366
      for (size_t i = 0; i < dup_op.size(); ++i) {
Y
Yang Yang(Tony) 已提交
367 368
        VLOG(10) << backward_descs[dup_op[i]]->Type() << " has " << out_name
                 << " duplicated";
F
Update  
fengjiayi 已提交
369
        std::string new_name = out_name + "@RENAME@" + std::to_string(i);
Y
Yang Yang(Tony) 已提交
370 371
        backward_descs[dup_op[i]]->RenameOutput(out_name, new_name);
        backward_descs[dup_op[i]]->RenameInput(out_name, next_g_name);
F
Update  
fengjiayi 已提交
372
        sum_op_inputs.emplace_back(new_name);
Y
Yang Yang(Tony) 已提交
373
        next_g_name = sum_op_inputs.back();
F
Update  
fengjiayi 已提交
374
      }
Y
Yu Yang 已提交
375 376 377
      std::unique_ptr<OpDesc> sum_op(new OpDesc("sum", {{"X", sum_op_inputs}},
                                                {{"Out", {out_name}}},
                                                AttributeMap{}));
F
fengjiayi 已提交
378
      pending_sum_ops.push_back({dup_op.back(), std::move(sum_op)});
F
Update  
fengjiayi 已提交
379 380
    }
  }
Y
Yang Yang(Tony) 已提交
381

Y
Yu Yang 已提交
382 383 384 385
  pending_sum_ops.sort([](const std::pair<size_t, std::unique_ptr<OpDesc>>& a,
                          const std::pair<size_t, std::unique_ptr<OpDesc>>& b) {
    return a.first > b.first;
  });
F
Update  
fengjiayi 已提交
386
  for (auto& p : pending_sum_ops) {
F
Update  
fengjiayi 已提交
387 388
    backward_descs.insert(backward_descs.begin() + p.first + 1,
                          std::move(p.second));
F
Update  
fengjiayi 已提交
389
  }
390

Y
Yang Yang(Tony) 已提交
391 392
  VLOG(5) << "MakeBlockBackward Finished";

F
fengjiayi 已提交
393 394 395
  return backward_descs;
}

Y
Yu Yang 已提交
396 397
static BlockDesc* CreateStepBlock(
    ProgramDesc& program_desc, std::unordered_set<std::string>* no_grad_vars,
Y
Yu Yang 已提交
398 399 400 401
    std::unordered_map<std::string, std::string>* grad_to_var,
    int step_block_idx) {
  auto backward_block_op_descs = MakeBlockBackward(program_desc, step_block_idx,
                                                   no_grad_vars, grad_to_var);
Y
Yu Yang 已提交
402
  BlockDesc* backward_block =
Y
Yu Yang 已提交
403 404 405 406 407 408 409
      program_desc.AppendBlock(*program_desc.MutableBlock(step_block_idx));
  for (auto& ptr : backward_block_op_descs) {
    backward_block->AppendAllocatedOp(move(ptr));
  }
  return backward_block;
}

Q
qiaolongfei 已提交
410
ParamGradInfoMap AppendBackward(
Y
Yu Yang 已提交
411
    ProgramDesc& program_desc, const VarDesc& target,
Q
qiaolongfei 已提交
412
    const std::unordered_set<std::string>& no_grad_vars) {
F
fengjiayi 已提交
413 414 415 416 417 418
  std::unordered_set<std::string> no_grad_var_names;
  no_grad_var_names.reserve(no_grad_vars.size() + 1);
  no_grad_var_names.insert(std::string(kEmptyVarName) + kGradVarSuffix);
  for (auto& name : no_grad_vars) {
    no_grad_var_names.insert(GradVarName(name));
  }
419

F
fengjiayi 已提交
420
  const int root_block_idx = 0;
421
  auto root_block = program_desc.MutableBlock(root_block_idx);
422 423

  std::string fill_one_op_out = GradVarName(target.Name());
F
fengjiayi 已提交
424
  bool is_scalar = target.GetShape() == std::vector<int64_t>{1};
425
  PADDLE_ENFORCE(is_scalar, "target should be scalar");
Y
Yu Yang 已提交
426 427
  VLOG(3) << "backward from loss=" << target.Name()
          << " data_type=" << target.GetDataType();
Y
Yu Yang 已提交
428 429 430 431 432
  std::unique_ptr<OpDesc> fill_one_op(
      new OpDesc("fill_constant", {}, {{"Out", {fill_one_op_out}}},
                 {{"shape", std::vector<int>{1}},
                  {"value", static_cast<float>(1.0)},
                  {"dtype", target.GetDataType()}}));
Q
QI JUN 已提交
433 434 435
  // infer var type of fill_one_op
  fill_one_op->InferVarType(root_block);

436 437
  root_block->AppendAllocatedOp(std::move(fill_one_op));
  size_t forward_op_num = root_block->OpSize();
438
  size_t forward_block_num = program_desc.Size();
Y
Yu Yang 已提交
439 440

  // Insert backward operators
441 442 443
  std::unordered_map<std::string, std::string> grad_to_var;
  auto backward_op_descs = MakeBlockBackward(program_desc, root_block_idx,
                                             &no_grad_var_names, &grad_to_var);
Y
Yu Yang 已提交
444

F
fengjiayi 已提交
445
  for (auto& ptr : backward_op_descs) {
446
    root_block->AppendAllocatedOp(std::move(ptr));
447
  }
Q
Qiao Longfei 已提交
448 449 450 451 452 453
  // Create Variable

  // Create target gradient variable
  std::unordered_map<std::string, GradVarInfo> retv;

  auto var = root_block->Var(fill_one_op_out);
Y
Yu Yang 已提交
454
  var->SetDataType(target.GetDataType());
F
fengjiayi 已提交
455
  var->SetShape(target.GetShape());
Q
Qiao Longfei 已提交
456 457 458 459
  auto& target_grad = retv[target.Name()];
  target_grad.name_ = fill_one_op_out;
  target_grad.block_idx_ = root_block_idx;
  target_grad.op_idx_ = static_cast<int>(forward_op_num);
460 461

  // create grad_var for all blocks in this program
462
  CreateGradVarInBlock(forward_op_num, grad_to_var, root_block, &retv);
463 464
  for (size_t block_index = forward_block_num;
       block_index < program_desc.Size(); ++block_index) {
465
    CreateGradVarInBlock(0, grad_to_var, program_desc.MutableBlock(block_index),
466
                         &retv);
F
fengjiayi 已提交
467
  }
Y
Yu Yang 已提交
468
  return retv;
F
Update  
fengjiayi 已提交
469 470
}

Y
Yu Yang 已提交
471 472
}  // namespace framework
}  // namespace paddle