From 89ffcec12e4c41f575e89e6edb503c32c838b8d2 Mon Sep 17 00:00:00 2001 From: Aston Zhang Date: Sat, 11 Aug 2018 21:29:34 +0000 Subject: [PATCH] add padding explanation --- chapter_natural-language-processing/embedding-training.md | 1 + 1 file changed, 1 insertion(+) diff --git a/chapter_natural-language-processing/embedding-training.md b/chapter_natural-language-processing/embedding-training.md index a70a506..7f84567 100644 --- a/chapter_natural-language-processing/embedding-training.md +++ b/chapter_natural-language-processing/embedding-training.md @@ -316,6 +316,7 @@ def train_embedding(num_epochs): # pred 形状:(batch_size, 1, max_len)。 pred = nd.batch_dot(emb_in, emb_out.swapaxes(1, 2)) # mask 和 label 形状:(batch_size, max_len)。 + # 避免填充对损失函数计算的影响。 l = (loss(pred.reshape(label.shape), label, mask) * mask.shape[1] / mask.sum(axis=1)) l.backward() -- GitLab