while_op.cc 20.0 KB
Newer Older
C
chengduo 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13
// Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
//     http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
Y
Yang Yang(Tony) 已提交
14

Y
Yi Wang 已提交
15 16 17
#include "paddle/fluid/framework/executor.h"
#include "paddle/fluid/framework/op_registry.h"
#include "paddle/fluid/framework/operator.h"
S
sneaxiy 已提交
18
#include "paddle/fluid/operators/controlflow/while_op_helper.h"
Y
Yang Yang(Tony) 已提交
19

W
wanghuancoder 已提交
20 21 22 23 24 25 26 27
namespace paddle {
namespace framework {
class InferShapeContext;
class OpDesc;
class VarDesc;
}  // namespace framework
}  // namespace paddle

Y
Yang Yang(Tony) 已提交
28 29 30 31 32 33
namespace paddle {
namespace operators {

using StepScopeVar = std::vector<framework::Scope *>;
using LoDTensor = framework::LoDTensor;

S
sneaxiy 已提交
34 35 36 37 38 39 40 41 42 43 44 45
namespace {  // NOLINT
static std::string GetSkipEagerDeletionVarsDebugString(
    const std::vector<std::string> &vars) {
  std::string str = "Skip " + std::to_string(vars.size()) +
                    " var(s) in eager deletion mode: ";
  for (auto &var : vars) {
    str.append(var);
    str.push_back(' ');
  }
  return str;
}
}  // NOLINT
Y
Yang Yang(Tony) 已提交
46 47 48 49 50 51 52 53

class WhileOp : public framework::OperatorBase {
 public:
  WhileOp(const std::string &type, const framework::VariableNameMap &inputs,
          const framework::VariableNameMap &outputs,
          const framework::AttributeMap &attrs)
      : framework::OperatorBase(type, inputs, outputs, attrs) {}

54 55 56
 private:
  void RunImpl(const framework::Scope &scope,
               const platform::Place &dev_place) const override {
57 58 59
    PADDLE_ENFORCE_NOT_NULL(scope.FindVar(Input(kCondition)),
                            platform::errors::NotFound(
                                "Input(Condition) of WhileOp is not found."));
60

Y
Yang Yang(Tony) 已提交
61
    auto &cond = scope.FindVar(Input(kCondition))->Get<LoDTensor>();
62 63 64 65 66 67
    PADDLE_ENFORCE_EQ(
        cond.dims(), paddle::framework::make_ddim({1}),
        platform::errors::InvalidArgument(
            "The shape of Input(Condition) of WhileOp must be 1. But now "
            "the Condition's shape is ",
            cond.dims().to_str(), ".\n"));
Y
Yang Yang(Tony) 已提交
68

D
dzhwinter 已提交
69
    framework::Executor executor(dev_place);
Y
Yu Yang 已提交
70
    auto *block = Attr<framework::BlockDesc *>(kStepBlock);
D
dzhwinter 已提交
71

Y
Yang Yang(Tony) 已提交
72 73 74 75
    auto *program = block->Program();

    auto step_scopes =
        scope.FindVar(Output(kStepScopes))->GetMutable<StepScopeVar>();
76 77 78 79 80 81 82 83 84 85 86

    if (step_scopes->size() > 0) {
      platform::DeviceContextPool::Instance().Get(dev_place)->Wait();
      for (auto &s : *step_scopes) {
        if (scope.HasKid(s)) {
          scope.DeleteScope(s);
        }
      }
      step_scopes->clear();
    }

87 88 89
    PADDLE_ENFORCE_EQ(step_scopes->size(), 0,
                      platform::errors::PreconditionNotMet(
                          "The Output(StepScope) of WhileOp should be empty."));
X
Xin Pan 已提交
90

91
    bool cond_data = GetCondData(cond);
C
chengduo 已提交
92
    bool is_test = Attr<bool>("is_test");
S
sneaxiy 已提交
93
    auto &skip_vars = Attr<std::vector<std::string>>(kSkipEagerDeletionVars);
S
sneaxiy 已提交
94
    VLOG(2) << GetSkipEagerDeletionVarsDebugString(skip_vars);
S
fix bug  
sneaxiy 已提交
95

S
sneaxiy 已提交
96
    auto ctx = executor.Prepare(*program, block->ID(), skip_vars);
97
    if (!is_test) {
98
      while (cond_data) {
99 100 101 102
        auto &current_scope = scope.NewScope();
        step_scopes->push_back(&current_scope);
        executor.RunPreparedContext(ctx.get(), &current_scope, false, true,
                                    true);
103 104
        cond_data =
            GetCondData(scope.FindVar(Input(kCondition))->Get<LoDTensor>());
105 106
      }
    } else {
Y
Yang Yang(Tony) 已提交
107
      auto &current_scope = scope.NewScope();
108
      executor.CreateVariables(*program, &current_scope, block->ID());
109
      while (cond_data) {
110 111 112 113 114 115 116 117 118 119 120 121 122 123 124
        for (auto &name : current_scope.LocalVarNames()) {
          auto *var = current_scope.Var(name);
          if (var->IsType<framework::LoDTensor>()) {
            // Clear all lod information for all lod_tensors.
            auto *t = var->GetMutable<framework::LoDTensor>();
            framework::LoD empty_lod;
            t->set_lod(empty_lod);
          } else if (var->IsType<framework::LoDTensorArray>()) {
            // Clear elements of all tensor arrays.
            auto *t = var->GetMutable<framework::LoDTensorArray>();
            t->clear();
          }
        }
        executor.RunPreparedContext(ctx.get(), &current_scope, false, false,
                                    false);
125 126
        cond_data =
            GetCondData(scope.FindVar(Input(kCondition))->Get<LoDTensor>());
C
chengduo 已提交
127
      }
128
      scope.DeleteScope(&current_scope);
Y
Yang Yang(Tony) 已提交
129 130 131 132 133 134
    }
  }
};

class WhileOpMaker : public framework::OpProtoAndCheckerMaker {
 public:
Y
Yu Yang 已提交
135
  void Make() override {
Y
Yang Yu 已提交
136
    AddInput(kX,
Y
Yang Yang(Tony) 已提交
137 138 139 140 141 142 143
             "A set of variables, which are required by operators inside the "
             "block of While Op.")
        .AsDuplicable();
    AddInput(
        kCondition,
        "(Bool) An scalar. When it's False, the While Op will be terminated.")
        .AsDuplicable();
Y
Yang Yang(Tony) 已提交
144
    AddOutput(kOutputs,
Y
Yang Yang(Tony) 已提交
145
              "A set of variables, which will be assigned with values "
Y
Yang Yang(Tony) 已提交
146
              "generated by the operators inside the block of While Op.")
Y
Yang Yang(Tony) 已提交
147 148 149 150 151
        .AsDuplicable();
    AddOutput(kStepScopes,
              "(StepScopeVar) A vector of local scope, which size equals the "
              "step number of While Op. The i'th scope storages temporary "
              "variables generated in the i'th step.");
Y
Yu Yang 已提交
152 153
    AddAttr<framework::BlockDesc *>(kStepBlock,
                                    "The step block inside WhileOp");
154 155 156 157
    AddAttr<bool>("is_test",
                  "(bool, default false) Set to true for inference only, false "
                  "for training. Some layers may run faster when this is true.")
        .SetDefault(false);
S
sneaxiy 已提交
158
    AddAttr<std::vector<std::string>>(kSkipEagerDeletionVars,
S
fix bug  
sneaxiy 已提交
159 160 161
                                      "Vars that would skip eager deletion."
                                      "Users should not set this manually.")
        .SetDefault(std::vector<std::string>());
Y
Yang Yang(Tony) 已提交
162 163 164 165 166 167 168 169 170 171 172 173
    AddComment(R"DOC(
)DOC");
  }
};

class WhileGradOp : public framework::OperatorBase {
 public:
  WhileGradOp(const std::string &type, const framework::VariableNameMap &inputs,
              const framework::VariableNameMap &outputs,
              const framework::AttributeMap &attrs)
      : framework::OperatorBase(type, inputs, outputs, attrs) {}

174 175 176
 private:
  void RunImpl(const framework::Scope &scope,
               const platform::Place &dev_place) const override {
177 178 179 180
    PADDLE_ENFORCE_EQ(
        Attr<bool>("is_test"), false,
        platform::errors::InvalidArgument(
            "WhileGradOp is only callable when is_test is false."));
181 182 183
    // get device context from pool
    platform::DeviceContextPool &pool = platform::DeviceContextPool::Instance();
    auto &dev_ctx = *pool.Get(dev_place);
D
dzhwinter 已提交
184
    framework::Executor executor(dev_place);
Y
Yu Yang 已提交
185
    auto *block = Attr<framework::BlockDesc *>(kStepBlock);
Y
Yang Yang(Tony) 已提交
186
    auto *program = block->Program();
S
sneaxiy 已提交
187 188

    auto &skip_vars = Attr<std::vector<std::string>>(kSkipEagerDeletionVars);
S
sneaxiy 已提交
189
    VLOG(2) << GetSkipEagerDeletionVarsDebugString(skip_vars);
S
sneaxiy 已提交
190
    auto ctx = executor.Prepare(*program, block->ID(), skip_vars);
Y
Yang Yang(Tony) 已提交
191 192 193 194

    auto *step_scopes =
        scope.FindVar(Input(kStepScopes))->GetMutable<StepScopeVar>();

Y
Yang Yang(Tony) 已提交
195 196 197 198
    auto outside_og_names = Inputs(framework::GradVarName(kOutputs));
    auto inside_og_names =
        Attr<std::vector<std::string>>("original_output_grad");

199 200 201 202 203 204 205 206
    PADDLE_ENFORCE_EQ(outside_og_names.size(), inside_og_names.size(),
                      platform::errors::InvalidArgument(
                          "The number of original output gradient names "
                          "does not match the number of backward input "
                          "gradient names. The number of Backward input "
                          "names is %d and the numbers of original output "
                          "gradient names is %d.",
                          outside_og_names.size(), inside_og_names.size()));
Y
Yang Yang(Tony) 已提交
207

Y
Yang Yang(Tony) 已提交
208 209
    for (auto cur_scope_iter = step_scopes->rbegin();
         cur_scope_iter != step_scopes->rend(); ++cur_scope_iter) {
M
minqiyang 已提交
210 211
      VLOG(3) << "Start backward at time_step "
              << cur_scope_iter - step_scopes->rbegin();
Y
Yang Yang(Tony) 已提交
212 213 214 215 216
      framework::Scope &cur_scope = **cur_scope_iter;
      // Link OG from outside to inside
      for (size_t i = 0; i < outside_og_names.size(); ++i) {
        auto outside_og_name = outside_og_names[i];
        auto inside_og_name = inside_og_names[i];
M
minqiyang 已提交
217 218
        VLOG(8) << "Linking outside " << outside_og_name << " --> inside "
                << inside_og_name;
C
chengduo 已提交
219 220 221 222
        if (scope.FindVar(outside_og_name) == nullptr) {
          continue;
        }

223 224
        auto &og_outside = *scope.FindVar(outside_og_name);
        auto &og_inside = *cur_scope.Var(inside_og_name);
S
sneaxiy 已提交
225
        if (og_outside.IsType<framework::LoDTensor>()) {
Y
Yang Yang(Tony) 已提交
226
          auto &outside_tensor = og_outside.Get<framework::LoDTensor>();
227
          auto &inside_tensor = *og_inside.GetMutable<framework::LoDTensor>();
Y
Yang Yang(Tony) 已提交
228 229
          inside_tensor.set_lod(outside_tensor.lod());
          inside_tensor.ShareDataWith(outside_tensor);
S
sneaxiy 已提交
230
        } else if (og_outside.IsType<framework::LoDTensorArray>()) {
231 232
          auto outside_array =
              og_outside.GetMutable<framework::LoDTensorArray>();
Y
Yang Yang(Tony) 已提交
233
          auto &inside_array =
234
              *og_inside.GetMutable<framework::LoDTensorArray>();
235 236 237
          inside_array.clear();
          inside_array.resize(outside_array->size());
          VLOG(8) << outside_og_name << " size = " << outside_array->size();
Y
Yang Yang(Tony) 已提交
238 239

          for (size_t j = 0; j < inside_array.size(); ++j) {
240 241 242 243 244 245 246
            if (!outside_array->at(j).IsInitialized()) {
              outside_array->at(j).Resize({0});
            }
            VLOG(8) << j << " " << outside_array->at(j).numel();
            if (outside_array->at(j).numel() != 0) {
              inside_array[j].set_lod(outside_array->at(j).lod());
              inside_array[j].ShareDataWith(outside_array->at(j));
Y
Yang Yang(Tony) 已提交
247
            } else {
248 249 250 251 252 253
              PADDLE_ENFORCE_EQ(
                  inside_array[j].numel(), 0,
                  platform::errors::InvalidArgument(
                      "The numel of %d-th element of var %s (LoDTensorArray) "
                      "in while block must be 0, but received its numel is %d.",
                      j, inside_og_name, inside_array[j].numel()));
Y
Yang Yang(Tony) 已提交
254 255
            }
          }
C
chengduo 已提交
256
        } else {
257 258 259
          PADDLE_THROW(platform::errors::Unimplemented(
              "Currently only support LoDTensor and LoDTensorArray in "
              "WhileGradOp."));
Y
Yang Yang(Tony) 已提交
260 261
        }
      }
C
chengduoZH 已提交
262 263
      executor.RunPreparedContext(ctx.get(), *cur_scope_iter, false, true,
                                  true);
Y
Yang Yang(Tony) 已提交
264

C
chengduo 已提交
265 266 267
      // The Outputs(kXGRAD) contains the names of the gradient of parameters
      // and inputs.
      auto &pg_ig_names = Outputs(kXGRAD);
Y
Yang Yu 已提交
268
      auto &p_names = Inputs(kX);
269 270 271 272 273 274 275
      PADDLE_ENFORCE_EQ(pg_ig_names.size(), p_names.size(),
                        platform::errors::PreconditionNotMet(
                            "The number of names in Outputs(X@GRAD) does not "
                            "match the number of names in Inputs(X). The "
                            "number of names in Outputs(X@GRAD) is %d and "
                            "the number of names in Inputs(X) is %d.",
                            pg_ig_names.size(), p_names.size()));
C
chengduo 已提交
276 277
      for (size_t param_id = 0; param_id < pg_ig_names.size(); ++param_id) {
        if (pg_ig_names[param_id] == framework::kEmptyVarName) {
278
          continue;  // parameter doesn't have gradient
Y
Yang Yang(Tony) 已提交
279 280
        }
        auto inside_grad_name = framework::GradVarName(p_names[param_id]);
Y
Yang Yang(Tony) 已提交
281

C
chengduo 已提交
282 283 284 285
        // for some grad_op, their input doesn't have gradient,
        // for example lookup_table_grad_op, the input(Idx) doesn't have
        // gradient.
        auto pg_ig_var = cur_scope.FindVar(inside_grad_name);
286 287 288
        PADDLE_ENFORCE_NOT_NULL(
            pg_ig_var, platform::errors::NotFound("Variable %s is not found.",
                                                  inside_grad_name));
C
chengduo 已提交
289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305
        if (pg_ig_var->IsType<framework::LoDTensorArray>()) {
          auto pg_ig_lod_t_arr =
              pg_ig_var->GetMutable<framework::LoDTensorArray>();
          bool empty = true;
          for (auto &each : *pg_ig_lod_t_arr) {
            if (each.numel() != 0) {
              empty = false;
              break;
            }
          }
          if (empty) {
            LOG(WARNING) << pg_ig_names[param_id]
                         << " is not found in cur_scope.";
            continue;
          }
        }

Y
Yang Yang(Tony) 已提交
306
        //  // TODO(tonyyang-svail): Not sure we need the following
Y
Yang Yang(Tony) 已提交
307 308 309 310 311 312 313 314 315 316 317
        //  // If does not compute gradient of that variable inside rnn,
        //  just
        //  // continue
        //  if (local_var_names.find(inside_grad_name) ==
        //  local_var_names.end()) {
        //    continue;
        //  }

        // zero gradient variable in step 0
        if (cur_scope_iter == step_scopes->rbegin()) {
          auto *var = (*cur_scope_iter)->FindVar(inside_grad_name);
318 319 320 321
          PADDLE_ENFORCE_NOT_NULL(
              var, platform::errors::NotFound("Variable %s is not found.",
                                              inside_grad_name));
          PADDLE_ENFORCE_EQ(
C
chengduoZH 已提交
322 323
              var->IsType<framework::LoDTensorArray>() ||
                  var->IsType<LoDTensor>(),
324 325 326 327
              true, platform::errors::InvalidArgument(
                        "Currently the type of var only can be LoDTensorArray, "
                        "or LoDTensor, but the received var[%s] is %s.",
                        inside_grad_name, framework::ToTypeName(var->Type())));
C
chengduo 已提交
328

Y
Yang Yang(Tony) 已提交
329 330 331
          if (var->IsType<LoDTensor>()) {
            auto &inside_tensor = var->Get<framework::LoDTensor>();
            framework::AttributeMap attrs;
Y
Yu Yang 已提交
332
            attrs["dtype"] = inside_tensor.type();
333
            attrs["shape"] = framework::vectorize<int>(inside_tensor.dims());
Y
Yang Yang(Tony) 已提交
334 335
            attrs["value"] = 0.0f;

C
chengduo 已提交
336
            auto var_name = pg_ig_names[param_id];
Y
Yang Yang(Tony) 已提交
337
            auto zero_op = framework::OpRegistry::CreateOp(
Y
Yiqun Liu 已提交
338
                "fill_constant", framework::VariableNameMap{},
339
                {{"Out", {var_name}}}, attrs);
D
dzhwinter 已提交
340
            zero_op->Run(scope, dev_place);
341 342 343
            scope.FindVar(var_name)
                ->GetMutable<framework::LoDTensor>()
                ->set_lod(inside_tensor.lod());
Y
Yang Yang(Tony) 已提交
344 345
          }
        }
Y
Yang Yang(Tony) 已提交
346
        auto new_inside_name = cur_scope.Rename(inside_grad_name);
Y
Yang Yang(Tony) 已提交
347
        auto sum_op = framework::OpRegistry::CreateOp(
C
chengduo 已提交
348 349
            "sum", {{"X", {pg_ig_names[param_id], new_inside_name}}},
            {{"Out", {pg_ig_names[param_id]}}},
350
            framework::AttributeMap{{"use_mkldnn", {false}}});
D
dzhwinter 已提交
351
        sum_op->Run(cur_scope, dev_place);
Y
Yang Yang(Tony) 已提交
352
        cur_scope.Rename(new_inside_name, inside_grad_name);
Y
Yang Yang(Tony) 已提交
353
      }
354 355
      dev_ctx.Wait();
      const_cast<framework::Scope &>(scope).DeleteScope(&cur_scope);
Y
Yang Yang(Tony) 已提交
356
    }
357
    step_scopes->clear();
Y
Yang Yang(Tony) 已提交
358 359 360
  }
};

H
hong 已提交
361 362
template <typename T>
class WhileGradOpMaker : public framework::SingleGradOpMaker<T> {
Y
Yang Yang(Tony) 已提交
363
 public:
H
hong 已提交
364
  using framework::SingleGradOpMaker<T>::SingleGradOpMaker;
Y
Yang Yang(Tony) 已提交
365 366

 protected:
367
  void Apply(GradOpPtr<T> while_grad) const override {
F
Update  
fengjiayi 已提交
368
    while_grad->SetType("while_grad");
H
hong 已提交
369 370 371
    while_grad->SetInput(kX, this->Input(kX));
    while_grad->SetInput(kOutputs, this->Output(kOutputs));
    while_grad->SetInput(kStepScopes, this->Output(kStepScopes));
F
Update  
fengjiayi 已提交
372 373

    auto *grad_block = this->grad_block_[0];
Y
Yu Yang 已提交
374 375
    auto *fwd_block = grad_block->ForwardBlock();
    auto *parent_block = grad_block->ParentBlock();
376 377 378

    // Not all of IGs will be generated by inner gradient operators of while op.
    // Ignore IGs that is not generated by the inside block.
F
Update  
fengjiayi 已提交
379 380 381 382
    std::unordered_set<std::string> inner_op_outputs;
    for (const auto *op : grad_block->AllOps()) {
      for (auto &oname : op->OutputArgumentNames()) {
        inner_op_outputs.insert(oname);
383 384
      }
    }
H
hong 已提交
385 386
    auto igs = this->InputGrad(kX, /*do not drop empty gradient*/ false);

387
    for (auto &each_ig : igs) {
F
Update  
fengjiayi 已提交
388
      if (inner_op_outputs.find(each_ig) == inner_op_outputs.end()) {
M
minqiyang 已提交
389
        VLOG(8) << "Ignore " << each_ig;
390 391 392
        each_ig = framework::kEmptyVarName;
      }
    }
F
Update  
fengjiayi 已提交
393
    while_grad->SetOutput(framework::GradVarName(kX), igs);
Y
Yang Yang(Tony) 已提交
394 395 396 397

    // OG should be re-calculated by step blocks, since many outputs of while op
    // do not need to calculate gradients.
    std::unordered_set<std::string> block_ins;
H
hong 已提交
398 399
    block_ins.reserve(this->Input(kX).size() + this->Output(kOutputs).size());
    for (auto &p : this->Input(kX)) {
F
fengjiayi 已提交
400 401
      block_ins.insert(p);
    }
H
hong 已提交
402
    for (auto &o : this->Output(kOutputs)) {
F
fengjiayi 已提交
403 404
      block_ins.insert(o);
    }
Y
Yu Yang 已提交
405
    std::unordered_set<std::string> output_grads;
F
Update  
fengjiayi 已提交
406 407 408 409
    for (const auto *op : grad_block->AllOps()) {
      for (auto &input_name : op->InputArgumentNames()) {
        // If the input of Op has been recorded or is generated by the forward
        // block, do not make it as input again.
Y
Yu Yang 已提交
410 411 412

        // The input is located in I/O or other op's outputs or the variable is
        // located in grad_block's parents
F
Update  
fengjiayi 已提交
413
        if (block_ins.find(input_name) != block_ins.end() ||
Y
Yu Yang 已提交
414 415
            (fwd_block->FindVarRecursive(input_name) != nullptr ||
             parent_block->FindVarRecursive(input_name) != nullptr)) {
Y
Yang Yang(Tony) 已提交
416 417
          continue;
        }
C
chengduo 已提交
418

Y
Yu Yang 已提交
419
        output_grads.insert(input_name);
Y
Yang Yang(Tony) 已提交
420
      }
F
Update  
fengjiayi 已提交
421
      for (auto &output_name : op->OutputArgumentNames()) {
Y
Yang Yang(Tony) 已提交
422
        block_ins.insert(output_name);
Y
Yang Yang(Tony) 已提交
423 424
      }
    }
Y
Yang Yang(Tony) 已提交
425

Y
Yu Yang 已提交
426 427 428 429 430
    std::vector<std::string> output_grads_list;
    output_grads_list.resize(output_grads.size());
    std::copy(output_grads.begin(), output_grads.end(),
              output_grads_list.begin());
    while_grad->SetInput(framework::GradVarName(kOutputs), output_grads_list);
F
Update  
fengjiayi 已提交
431 432

    while_grad->SetAttrMap(this->Attrs());
A
Abhinav Arora 已提交
433
    while_grad->SetBlockAttr(kStepBlock, grad_block);
Y
Yang Yang(Tony) 已提交
434 435
    // record the original output gradient names, since the gradient name of
    // while operator could be renamed.
Y
Yu Yang 已提交
436
    while_grad->SetAttr("original_output_grad", output_grads_list);
Y
Yang Yang(Tony) 已提交
437

S
sneaxiy 已提交
438
    while_grad->SetAttr(kSkipEagerDeletionVars, std::vector<std::string>());
Y
Yang Yang(Tony) 已提交
439 440 441
  }
};

442 443
class WhileGradOpVarTypeInference
    : public framework::StaticGraphVarTypeInference {
Y
Yang Yang(Tony) 已提交
444
 public:
M
minqiyang 已提交
445
  void operator()(framework::InferVarTypeContext *ctx) const override {
446 447
    auto p_names = Input(ctx, kX);
    auto pg_ig_names = Output(ctx, framework::GradVarName(kX));
Y
Yang Yang(Tony) 已提交
448 449

    for (size_t i = 0; i < p_names.size(); ++i) {
450
      if (HasVar(ctx, pg_ig_names[i])) {
M
minqiyang 已提交
451
        VLOG(5) << "Setting " << pg_ig_names[i] << " following " << p_names[i]
452 453 454
                << " type: " << GetType(ctx, p_names[i]);
        SetType(ctx, pg_ig_names[i], GetType(ctx, p_names[i]));
        SetDataType(ctx, pg_ig_names[i], GetDataType(ctx, p_names[i]));
Y
Yang Yang(Tony) 已提交
455 456 457 458 459 460 461 462
      }
    }
  }
};

class WhileGradOpShapeInference : public framework::InferShapeBase {
 public:
  void operator()(framework::InferShapeContext *ctx) const override {
Y
Yang Yu 已提交
463 464
    ctx->HasInputs(kX);
    ctx->HasOutputs(framework::GradVarName(kX));
Y
Yang Yang(Tony) 已提交
465 466
    ctx->HasInputs(kOutputs);
    ctx->HasInputs(framework::GradVarName(kOutputs));
C
chengduo 已提交
467
    auto pg_ig_names = ctx->Outputs(kXGRAD);
X
Xin Pan 已提交
468 469 470 471
    std::vector<framework::InferShapeVarPtr> in_var_ptrs =
        ctx->GetInputVarPtrs(kX);
    std::vector<framework::InferShapeVarPtr> out_var_ptrs =
        ctx->GetOutputVarPtrs(kXGRAD);
472 473 474 475
    PADDLE_ENFORCE_EQ(in_var_ptrs.size(), out_var_ptrs.size(),
                      platform::errors::InvalidArgument(
                          "The size of Inputs(X) must be the same as "
                          "the size of Outputs(X@GRAD)."));
X
Xin Pan 已提交
476 477

    for (size_t i = 0; i < in_var_ptrs.size(); ++i) {
C
chengduo 已提交
478
      if (pg_ig_names[i] == framework::kEmptyVarName) {
Y
Yang Yang(Tony) 已提交
479 480
        continue;
      }
481
      framework::VarDesc *in_var =
482 483
          BOOST_GET(framework::VarDesc *, in_var_ptrs[i]);
      BOOST_GET(framework::VarDesc *, out_var_ptrs[i])
484
          ->SetShape(in_var->GetShape());
Y
Yang Yang(Tony) 已提交
485 486 487 488
    }
  }
};

Y
Yang Yang(Tony) 已提交
489 490 491
}  // namespace operators
}  // namespace paddle

H
hong 已提交
492 493 494
REGISTER_OPERATOR(
    while, paddle::operators::WhileOp, paddle::operators::WhileOpMaker,
    paddle::operators::WhileGradOpMaker<paddle::framework::OpDesc>);
Y
Yang Yang(Tony) 已提交
495 496 497
REGISTER_OPERATOR(while_grad, paddle::operators::WhileGradOp,
                  paddle::operators::WhileGradOpShapeInference,
                  paddle::operators::WhileGradOpVarTypeInference);