diff --git a/word2vec/README.en.md b/word2vec/README.en.md index b9729fce7cb58546f64260ddc5f53962c50c558c..e4ce6e5126432830d894305ada52ad84a4ed972f 100644 --- a/word2vec/README.en.md +++ b/word2vec/README.en.md @@ -103,7 +103,7 @@ where $f(w_t, w_{t-1}, ..., w_{t-n+1})$ represents the conditional probability o (Translation of words in figure 2: 图2文字翻译 - 输入:Input; -- 全连接:Fully-Connection Layer +- 全连接:Fully-Connected Layer - 词向量:Word Embedding - 词向量连接:Word Embedding Concatenation - 分类:Classification @@ -172,6 +172,18 @@ As illustrated in the figure above, skip-gram model maps the word embedding of t ## Data Preparation ## Model Configuration +
+
+ Figure 5. N-gram neural network model in model configuration
+