model.py 4.7 KB
Newer Older
T
tangwei 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14
#   Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

X
xujiaqi01 已提交
15 16 17 18 19
import paddle.fluid as fluid
import paddle.fluid.layers.nn as nn
import paddle.fluid.layers.tensor as tensor
import paddle.fluid.layers.control_flow as cf

T
tangwei 已提交
20
from paddlerec.core.model import Model as ModelBase
T
for mat  
tangwei 已提交
21
from paddlerec.core.utils import envs
T
tangwei 已提交
22 23


X
xujiaqi01 已提交
24 25 26 27 28
class Model(ModelBase):
    def __init__(self, config):
        ModelBase.__init__(self, config)
        self.cost = None
        self.metrics = {}
T
tangwei 已提交
29 30 31 32
        self.vocab_text_size = envs.get_global_env("vocab_text_size", None,
                                                   self._namespace)
        self.vocab_tag_size = envs.get_global_env("vocab_tag_size", None,
                                                  self._namespace)
T
for mat  
tangwei 已提交
33 34 35 36 37
        self.emb_dim = envs.get_global_env("emb_dim", None, self._namespace)
        self.hid_dim = envs.get_global_env("hid_dim", None, self._namespace)
        self.win_size = envs.get_global_env("win_size", None, self._namespace)
        self.margin = envs.get_global_env("margin", None, self._namespace)
        self.neg_size = envs.get_global_env("neg_size", None, self._namespace)
X
xujiaqi01 已提交
38 39

    def train_net(self):
T
tangwei 已提交
40 41 42
        """ network"""
        text = fluid.data(
            name="text", shape=[None, 1], lod_level=1, dtype='int64')
X
xujiaqi01 已提交
43 44 45 46 47 48 49 50
        pos_tag = fluid.data(
            name="pos_tag", shape=[None, 1], lod_level=1, dtype='int64')
        neg_tag = fluid.data(
            name="neg_tag", shape=[None, 1], lod_level=1, dtype='int64')

        self._data_var = [text, pos_tag, neg_tag]

        text_emb = fluid.embedding(
T
tangwei 已提交
51 52 53
            input=text,
            size=[self.vocab_text_size, self.emb_dim],
            param_attr="text_emb")
X
xujiaqi01 已提交
54 55
        text_emb = fluid.layers.squeeze(input=text_emb, axes=[1])
        pos_tag_emb = fluid.embedding(
T
tangwei 已提交
56 57 58
            input=pos_tag,
            size=[self.vocab_tag_size, self.emb_dim],
            param_attr="tag_emb")
X
xujiaqi01 已提交
59 60
        pos_tag_emb = fluid.layers.squeeze(input=pos_tag_emb, axes=[1])
        neg_tag_emb = fluid.embedding(
T
tangwei 已提交
61 62 63
            input=neg_tag,
            size=[self.vocab_tag_size, self.emb_dim],
            param_attr="tag_emb")
X
xujiaqi01 已提交
64 65 66 67 68 69 70 71 72 73 74 75 76
        neg_tag_emb = fluid.layers.squeeze(input=neg_tag_emb, axes=[1])

        conv_1d = fluid.nets.sequence_conv_pool(
            input=text_emb,
            num_filters=self.hid_dim,
            filter_size=self.win_size,
            act="tanh",
            pool_type="max",
            param_attr="cnn")
        text_hid = fluid.layers.fc(input=conv_1d,
                                   size=self.emb_dim,
                                   param_attr="text_hid")
        cos_pos = nn.cos_sim(pos_tag_emb, text_hid)
T
tangwei 已提交
77 78
        mul_text_hid = fluid.layers.sequence_expand_as(
            x=text_hid, y=neg_tag_emb)
X
xujiaqi01 已提交
79 80 81 82 83 84 85 86
        mul_cos_neg = nn.cos_sim(neg_tag_emb, mul_text_hid)
        cos_neg_all = fluid.layers.sequence_reshape(
            input=mul_cos_neg, new_dim=self.neg_size)
        #choose max negtive cosine
        cos_neg = nn.reduce_max(cos_neg_all, dim=1, keep_dim=True)
        #calculate hinge loss
        loss_part1 = nn.elementwise_sub(
            tensor.fill_constant_batch_size_like(
T
tangwei 已提交
87 88 89 90
                input=cos_pos,
                shape=[-1, 1],
                value=self.margin,
                dtype='float32'),
X
xujiaqi01 已提交
91 92 93 94 95 96 97 98 99 100
            cos_pos)
        loss_part2 = nn.elementwise_add(loss_part1, cos_neg)
        loss_part3 = nn.elementwise_max(
            tensor.fill_constant_batch_size_like(
                input=loss_part2, shape=[-1, 1], value=0.0, dtype='float32'),
            loss_part2)
        avg_cost = nn.mean(loss_part3)
        less = tensor.cast(cf.less_than(cos_neg, cos_pos), dtype='float32')
        correct = nn.reduce_sum(less)
        self.cost = avg_cost
T
tangwei 已提交
101

X
xujiaqi01 已提交
102 103 104
        self.metrics["correct"] = correct
        self.metrics["cos_pos"] = cos_pos

T
tangwei 已提交
105
    def get_avg_cost(self):
X
xujiaqi01 已提交
106 107 108 109 110 111
        return self.cost

    def get_metrics(self):
        return self.metrics

    def optimizer(self):
T
tangwei 已提交
112 113
        learning_rate = envs.get_global_env("hyper_parameters.base_lr", None,
                                            self._namespace)
X
xujiaqi01 已提交
114 115 116 117 118
        sgd_optimizer = fluid.optimizer.Adagrad(learning_rate=learning_rate)
        return sgd_optimizer

    def infer_net(self, parameter_list):
        self.train_net()