提交 ce7bbecd 编写于 作者: W wangxiao

add softmax

上级 0955e2d8
...@@ -75,8 +75,9 @@ class TaskParadigm(task_paradigm): ...@@ -75,8 +75,9 @@ class TaskParadigm(task_paradigm):
name=scope_name+"cls_out_b", initializer=fluid.initializer.Constant(0.))) name=scope_name+"cls_out_b", initializer=fluid.initializer.Constant(0.)))
if self._is_training: if self._is_training:
inputs = fluid.layers.softmax(logits)
loss = fluid.layers.cross_entropy( loss = fluid.layers.cross_entropy(
input=logits, label=label_ids) input=inputs, label=label_ids)
loss = layers.mean(loss) loss = layers.mean(loss)
return {"loss": loss} return {"loss": loss}
else: else:
......
...@@ -79,8 +79,9 @@ class TaskParadigm(task_paradigm): ...@@ -79,8 +79,9 @@ class TaskParadigm(task_paradigm):
initializer=fluid.initializer.Constant(0.))) initializer=fluid.initializer.Constant(0.)))
if self._is_training: if self._is_training:
inputs = fluid.layers.softmax(logits)
ce_loss = fluid.layers.cross_entropy( ce_loss = fluid.layers.cross_entropy(
input=logits, label=labels) input=inputs, label=labels)
loss = fluid.layers.mean(x=ce_loss) loss = fluid.layers.mean(x=ce_loss)
return {'loss': loss} return {'loss': loss}
else: else:
......
...@@ -100,8 +100,9 @@ class TaskParadigm(task_paradigm): ...@@ -100,8 +100,9 @@ class TaskParadigm(task_paradigm):
is_bias=True) is_bias=True)
if self._is_training: if self._is_training:
inputs = fluid.layers.softmax(fc_out)
mask_lm_loss = fluid.layers.cross_entropy( mask_lm_loss = fluid.layers.cross_entropy(
input=fc_out, label=mask_label) input=inputs, label=mask_label)
loss = fluid.layers.mean(mask_lm_loss) loss = fluid.layers.mean(mask_lm_loss)
return {'loss': loss} return {'loss': loss}
else: else:
......
...@@ -102,8 +102,9 @@ class TaskParadigm(task_paradigm): ...@@ -102,8 +102,9 @@ class TaskParadigm(task_paradigm):
def _compute_single_loss(logits, positions): def _compute_single_loss(logits, positions):
"""Compute start/en """Compute start/en
d loss for mrc model""" d loss for mrc model"""
inputs = fluid.layers.softmax(logits)
loss = fluid.layers.cross_entropy( loss = fluid.layers.cross_entropy(
input=logits, label=positions) input=inputs, label=positions)
loss = fluid.layers.mean(x=loss) loss = fluid.layers.mean(x=loss)
return loss return loss
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册