From e3161bb61a4686d96588bc1eb86c3edc0e26e6ee Mon Sep 17 00:00:00 2001 From: Yang Yang Date: Tue, 10 Oct 2017 04:49:45 +0000 Subject: [PATCH] pass simple backward --- paddle/framework/executor_test.cc | 51 ++++++++++++++++++------------- 1 file changed, 29 insertions(+), 22 deletions(-) diff --git a/paddle/framework/executor_test.cc b/paddle/framework/executor_test.cc index 0515fb221..9f8a6f859 100644 --- a/paddle/framework/executor_test.cc +++ b/paddle/framework/executor_test.cc @@ -32,6 +32,8 @@ USE_OP(fetch); USE_OP(mul); USE_OP(sum); USE_OP(squared_l2_distance); +USE_OP(fill_constant); +USE_OP(sgd); using std::string; using namespace paddle::platform; @@ -124,6 +126,7 @@ class ExecutorTesterRandom : public ::testing::Test { // flush init_program.Proto(); + // run block auto temp_root_block = pdesc_.add_blocks(); temp_root_block->set_idx(0); temp_root_block->set_parent_idx(-1); @@ -131,6 +134,7 @@ class ExecutorTesterRandom : public ::testing::Test { paddle::framework::ProgramDescBind::Instance(&pdesc_); paddle::framework::BlockDescBind* root_block = program.Block(0); + // forward AddOp("gaussian_random", {}, {{"Out", {"a"}}}, {{"dims", std::vector{batch_size, input_dim}}}, root_block); AddOp("mul", {{"X", {"a"}}, {"Y", {"w1"}}}, {{"Out", {"b"}}}, {}, @@ -141,30 +145,33 @@ class ExecutorTesterRandom : public ::testing::Test { {{"Out", {"l2_distance"}}, {"sub_result", {"l2_distance_sub"}}}, {}, root_block); - // AddOp("gaussian_random", {}, {{"Out", {"l2_distance@GRAD"}}}, - // {{"dims", std::vector{batch_size, 1}}}, root_block); - // AppendBackward(program, {}); - - // program.Proto(); - - // for (auto& op : pdesc_.blocks(0).ops()) { - // if (op.type() == "sum") { - // LOG(INFO) << "Here"; - // for (auto& var : op.inputs()) { - // for (auto& argu : var.arguments()) { - // LOG(INFO) << var.parameter() << " " << argu; - // } - // } - // } - // } - - AddOp("fetch", {{"Input", {"l2_distance"}}}, {}, - {{"dims", std::vector{batch_size}}, {"col", 1}}, root_block); + // backward + AddOp("fill_constant", {}, {{"Out", {"l2_distance@GRAD"}}}, + {{"shape", std::vector{batch_size, 1}}, {"value", float(1.0)}}, + root_block); + AppendBackward(program, {}); + + // update + AddOp("fill_constant", {}, {{"Out", {"learning_rate"}}}, + {{"shape", std::vector{1}}, {"value", float(1.0)}}, root_block); + AddOp("sgd", {{"Param", {"w1"}}, + {"LearningRate", {"learning_rate"}}, + {"Grad", {"w1@GRAD"}}}, + {{"ParamOut", {"w1"}}}, {}, root_block); + AddOp("sgd", {{"Param", {"w2"}}, + {"LearningRate", {"learning_rate"}}, + {"Grad", {"w2@GRAD"}}}, + {{"ParamOut", {"w2"}}}, {}, root_block); + + AddOp("fetch", {{"Input", {"w1"}}}, {}, + {{"dims", std::vector{input_dim, embed_dim}}, {"col", 0}}, + root_block); + AddOp("fetch", {{"Input", {"w2"}}}, {}, + {{"dims", std::vector{embed_dim, input_dim}}, {"col", 1}}, + root_block); + // flush program.Proto(); - - // TODO(tonyyang-svail): - // - Test with Backward } protected: -- GitLab