From 49c31febb5fb5073705d2d3b17b8954ecea10ec1 Mon Sep 17 00:00:00 2001 From: tensor-tang Date: Mon, 27 Aug 2018 17:06:18 +0800 Subject: [PATCH] fix typo and op test --- .../operators/fusion_seqexpand_concat_fc_op.cc | 13 ++++++------- .../unittests/test_fusion_seqexpand_concat_fc_op.py | 8 ++++---- 2 files changed, 10 insertions(+), 11 deletions(-) diff --git a/paddle/fluid/operators/fusion_seqexpand_concat_fc_op.cc b/paddle/fluid/operators/fusion_seqexpand_concat_fc_op.cc index 641851585d..90aba5fe89 100644 --- a/paddle/fluid/operators/fusion_seqexpand_concat_fc_op.cc +++ b/paddle/fluid/operators/fusion_seqexpand_concat_fc_op.cc @@ -63,7 +63,7 @@ void FusionSeqExpandConcatFCOp::InferShape( framework::OpKernelType FusionSeqExpandConcatFCOp::GetExpectedKernelType( const framework::ExecutionContext& ctx) const { return framework::OpKernelType( - framework::ToDataType(ctx.Input("X")->type()), + framework::ToDataType(ctx.MultiInput("X")[0]->type()), ctx.device_context()); } @@ -113,7 +113,7 @@ class FusionSeqExpandConcatFCOpKernel : public framework::OpKernel { auto* w = ctx.Input("FCWeight"); auto* b = ctx.Input("FCBias"); auto* out = ctx.Output("Out"); - auto* fc_out = ctx.Output("FCOUT"); + auto* fc_out = ctx.Output("FCOut"); auto* ref_in = ins[0]; auto ref_lod = ref_in->lod(); @@ -164,7 +164,7 @@ class FusionSeqExpandConcatFCOpKernel : public framework::OpKernel { math::FCCompute(blas, total_T, D, M0, ref_in_data, w_data, out_data, b ? b->data() : NULL); w_data = w_data + M0 * D; - // first one use write on + // first write on blas.MatMul(N, D, M1, in1_data, w_data, fc_out_data); w_data = w_data + M1 * D; for (size_t i = 2; i < ins.size(); ++i) { @@ -175,16 +175,15 @@ class FusionSeqExpandConcatFCOpKernel : public framework::OpKernel { K, w_data, D, static_cast(1), fc_out_data, D); w_data = w_data + K * D; } - + T* cur_out_data = out_data; for (int i = 0; i < N; ++i) { int seq_len = ref_lod[0][i + 1] - ref_lod[0][i]; T* src = fc_out_data + i * D; for (int step = 0; step < seq_len; ++step) { - blas.VADD(D, out_data, src, out_data); - out_data = out_data + D; + blas.VADD(D, cur_out_data, src, cur_out_data); + cur_out_data = cur_out_data + D; } } - fc_act(total_T * D, out_data, out_data); } }; diff --git a/python/paddle/fluid/tests/unittests/test_fusion_seqexpand_concat_fc_op.py b/python/paddle/fluid/tests/unittests/test_fusion_seqexpand_concat_fc_op.py index 7baf39eb3f..aeee3a9999 100644 --- a/python/paddle/fluid/tests/unittests/test_fusion_seqexpand_concat_fc_op.py +++ b/python/paddle/fluid/tests/unittests/test_fusion_seqexpand_concat_fc_op.py @@ -80,16 +80,16 @@ class TestFusionSeqExpandConcatFCOp(OpTest): out = fusion_seqexpand_concat_fc(xs, self.lod, w, b, ACTIVATION[self.fc_act]) - self.inputs = {'X': [(x0, self.lod)], 'FCWeight': w} - normal_lod = [i for i in range(bs + 1)] + self.inputs = {'X': [('x0', (x0, self.lod))], 'FCWeight': w} + normal_lod = [[1] * bs] for i in range(num_inputs - 1): - self.inputs['X'].append((xs[i + 1], normal_lod)) + self.inputs['X'].append(('x%d' % (i + 1), (xs[i + 1], normal_lod))) if self.with_bias: self.inputs['FCBias'] = b self.outputs = {'Out': (out, self.lod)} - self.attrs = {'fc_activation': self.fc_act, } + self.attrs = {'fc_activation': self.fc_act} def test_check_output(self): self.check_output() -- GitLab