#include #include #include #include USE_OP(add_two); USE_OP(mul); USE_OP(sigmoid); USE_OP(softmax); namespace paddle { namespace framework { static int infer_shape_cnt = 0; static int run_cnt = 0; class TestOp : public OperatorBase { public: void InferShape(const framework::Scope& scope) const override { ++infer_shape_cnt; } void Run(const framework::Scope& scope, const paddle::platform::DeviceContext& dev_ctx) const override { ++run_cnt; } }; template void AssertSameVectorWithoutOrder(const std::vector& expected, const std::vector& actual) { ASSERT_EQ(expected.size(), actual.size()); std::unordered_set expected_set; for (auto& tmp : expected) { expected_set.insert(tmp); } for (auto& act : actual) { ASSERT_NE(expected_set.end(), expected_set.find(act)); } } TEST(OpKernel, all) { auto net = std::make_shared(); ASSERT_NE(net, nullptr); auto op1 = std::make_shared(); op1->inputs_ = {"x", "w1", "b1"}; op1->outputs_ = {"y"}; net->AddOp(op1); auto op2 = std::make_shared(); op2->inputs_ = {"y", "w2", "b2"}; op2->outputs_ = {"z"}; net->AddOp(op2); net->CompleteAddOp(); AssertSameVectorWithoutOrder({"x", "w1", "b1", "w2", "b2"}, net->inputs_); AssertSameVectorWithoutOrder({"y", "z"}, net->outputs_); auto tmp_idx_iter = net->attrs_.find("temporary_index"); ASSERT_NE(net->attrs_.end(), tmp_idx_iter); auto& tmp_idx = boost::get>(tmp_idx_iter->second); ASSERT_EQ(1UL, tmp_idx.size()); ASSERT_EQ("y", net->outputs_[tmp_idx[0]]); Scope scope; platform::CPUDeviceContext dev_ctx; net->InferShape(scope); net->Run(scope, dev_ctx); ASSERT_EQ(2, infer_shape_cnt); ASSERT_EQ(2, run_cnt); ASSERT_THROW(net->AddOp(op2), paddle::platform::EnforceNotMet); } //! TODO(yuyang18): Refine Backward Op. // TEST(AddBackwardOp, TestGradOp) { // auto net = std::make_shared(); // ASSERT_NE(net, nullptr); // net->AddOp(framework::OpRegistry::CreateOp("mul", {"X", "Y"}, {"Out"}, {})); // net->AddOp( // framework::OpRegistry::CreateOp("add_two", {"X", "Y"}, {"Out"}, {})); // net->AddOp(framework::OpRegistry::CreateOp("add_two", {"X", "Y"}, {""}, // {})); // auto grad_ops = AddBackwardOp(net); // for (auto& op : grad_ops->ops_) { // op->DebugString(); // } //} } // namespace framework } // namespace paddle