From 935ceaa1a9f14d3b8b6e98ca134a6ff7c2a4252c Mon Sep 17 00:00:00 2001 From: Yibing Liu Date: Tue, 5 Mar 2019 12:44:07 +0800 Subject: [PATCH] Add bert to the models guide (#1830) --- .gitmodules | 3 +++ README.md | 1 + fluid/PaddleNLP/LARK | 1 + 3 files changed, 5 insertions(+) create mode 160000 fluid/PaddleNLP/LARK diff --git a/.gitmodules b/.gitmodules index 4280afc1..d0281b7c 100644 --- a/.gitmodules +++ b/.gitmodules @@ -7,3 +7,6 @@ [submodule "fluid/PaddleNLP/Senta"] path = fluid/PaddleNLP/Senta url = https://github.com/baidu/Senta.git +[submodule "fluid/PaddleNLP/LARK"] + path = fluid/PaddleNLP/LARK + url = https://github.com/PaddlePaddle/LARK diff --git a/README.md b/README.md index 98cef358..182e6252 100644 --- a/README.md +++ b/README.md @@ -43,6 +43,7 @@ PaddlePaddle 提供了丰富的计算单元,使得用户可以采用模块化 模型|简介|模型优势|参考论文 --|:--:|:--:|:--: [Transformer](./fluid/PaddleNLP/neural_machine_translation/transformer/README_cn.md)|机器翻译模型|基于self-attention,计算复杂度小,并行度高,容易学习长程依赖,翻译效果更好|[Attention Is All You Need](https://arxiv.org/abs/1706.03762) +[BERT](https://github.com/PaddlePaddle/LARK/tree/develop/BERT)|语义表示模型|在多个 NLP 任务上取得 SOTA 效果,支持多卡多机训练,支持混合精度训练|[BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding](https://arxiv.org/abs/1810.04805) [LAC](https://github.com/baidu/lac/blob/master/README.md)|联合的词法分析模型|能够整体性地完成中文分词、词性标注、专名识别任务|[Chinese Lexical Analysis with Deep Bi-GRU-CRF Network](https://arxiv.org/abs/1807.01882) [Senta](https://github.com/baidu/Senta/blob/master/README.md)|情感倾向分析模型集|百度AI开放平台中情感倾向分析模型|- [DAM](./fluid/PaddleNLP/deep_attention_matching_net)|语义匹配模型|百度自然语言处理部发表于ACL-2018的工作,用于检索式聊天机器人多轮对话中应答的选择|[Multi-Turn Response Selection for Chatbots with Deep Attention Matching Network](http://aclweb.org/anthology/P18-1103) diff --git a/fluid/PaddleNLP/LARK b/fluid/PaddleNLP/LARK new file mode 160000 index 00000000..8dbdf489 --- /dev/null +++ b/fluid/PaddleNLP/LARK @@ -0,0 +1 @@ +Subproject commit 8dbdf4892a9c22a39a20537fd8584b760f41d963 -- GitLab