Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
我傻x
bert
提交
68d50765
B
bert
项目概览
我傻x
/
bert
与 Fork 源项目一致
从无法访问的项目Fork
通知
2
Star
1
Fork
1
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
B
bert
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
68d50765
编写于
11月 24, 2018
作者:
J
Jacob Devlin
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Adding BERT-Large cased
上级
32ac4763
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
10 addition
and
2 deletion
+10
-2
README.md
README.md
+10
-2
未找到文件。
README.md
浏览文件 @
68d50765
...
...
@@ -212,6 +212,10 @@ These models are all released under the same license as the source code (Apache
For information about the Multilingual and Chinese model, see the
[
Multilingual README
](
https://github.com/google-research/bert/blob/master/multilingual.md
)
.
**
When using a cased model, make sure to pass
`--do_lower=False`
to the training
scripts. (Or pass
`do_lower_case=False`
directly to
`FullTokenizer`
if you're
using your own script.)
**
The links to the models are here (right-click, 'Save link as...' on the name):
*
**[`BERT-Base, Uncased`](https://storage.googleapis.com/bert_models/2018_10_18/uncased_L-12_H-768_A-12.zip)**
:
...
...
@@ -220,8 +224,8 @@ The links to the models are here (right-click, 'Save link as...' on the name):
24-layer, 1024-hidden, 16-heads, 340M parameters
*
**[`BERT-Base, Cased`](https://storage.googleapis.com/bert_models/2018_10_18/cased_L-12_H-768_A-12.zip)**
:
12-layer, 768-hidden, 12-heads , 110M parameters
*
**
`BERT-Large, Cased`**
: 24-layer, 1024-hidden, 16-heads, 340M parameters
(Not available yet. Needs to be re-generated).
*
**
[`BERT-Large, Cased`](https://storage.googleapis.com/bert_models/2018_10_18/cased_L-24_H-1024_A-16.zip)**
:
24-layer, 1024-hidden, 16-heads, 340M parameters
*
**[`BERT-Base, Multilingual Cased (New, recommended)`](https://storage.googleapis.com/bert_models/2018_11_23/multi_cased_L-12_H-768_A-12.zip)**
:
104 languages, 12-layer, 768-hidden, 12-heads, 110M parameters
*
**[`BERT-Base, Multilingual Uncased (Orig, not recommended)`](https://storage.googleapis.com/bert_models/2018_11_03/multilingual_L-12_H-768_A-12.zip)**
:
...
...
@@ -826,6 +830,10 @@ accuracy numbers.
### Pre-training tips and caveats
*
**
If using your own vocabulary, make sure to change
`vocab_size`
in
`bert_config.json`
. If you use a larger vocabulary without changing this,
you will likely get NaNs when training on GPU or TPU due to unchecked
out-of-bounds access.
**
*
If your task has a large domain-specific corpus available (e.g., "movie
reviews" or "scientific papers"), it will likely be beneficial to run
additional steps of pre-training on your corpus, starting from the BERT
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录