diff --git a/python/paddle/trainer_config_helpers/networks.py b/python/paddle/trainer_config_helpers/networks.py index 120c9d11a5ebaa72b94590e596fd4362c552f979..3821d075cba5d39b5808a39093b8570d9302b667 100644 --- a/python/paddle/trainer_config_helpers/networks.py +++ b/python/paddle/trainer_config_helpers/networks.py @@ -1457,11 +1457,13 @@ def dot_product_attention(encoded_sequence, expanded = expand_layer( input=transformed_state, - expanded_as=encoded_sequence, + expand_as=encoded_sequence, name='%s_expand' % name) m = linear_comb_layer( - weights=expanded, vectors=encoded_sequence, name='%s_dot-product') + weights=expanded, + vectors=encoded_sequence, + name='%s_dot-product' % name) attention_weight = fc_layer( input=m,