Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
DeepSpeech
提交
d31dbabb
D
DeepSpeech
项目概览
PaddlePaddle
/
DeepSpeech
大约 2 年 前同步成功
通知
210
Star
8425
Fork
1598
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
245
列表
看板
标记
里程碑
合并请求
3
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
D
DeepSpeech
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
245
Issue
245
列表
看板
标记
里程碑
合并请求
3
合并请求
3
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
d31dbabb
编写于
1月 12, 2022
作者:
J
Jackwaterveg
提交者:
GitHub
1月 12, 2022
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
[Doc] Update released model for r0.1.1 (#1316)
* updated ASR released model * update the doc, test=doc_fix
上级
8f507ba4
变更
1
隐藏空白更改
内联
并排
Showing
1 changed file
with
8 addition
and
9 deletion
+8
-9
docs/source/released_model.md
docs/source/released_model.md
+8
-9
未找到文件。
docs/source/released_model.md
浏览文件 @
d31dbabb
...
...
@@ -5,14 +5,13 @@
### Speech Recognition Model
Acoustic Model | Training Data | Token-based | Size | Descriptions | CER | WER | Hours of speech | Example Link
:-------------:| :------------:| :-----: | -----: | :-----: |:-----:| :-----: | :-----: | :-----:
[
Ds2 Online Aishell ASR0 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr0/aishell_ds2_online_cer8.00_release.tar.gz
)
| Aishell Dataset | Char-based | 345 MB | 2 Conv + 5 LSTM layers with only forward direction | 0.080 |-| 151 h |
[
D2 Online Aishell ASR0
](
../../examples/aishell/asr0
)
[
Ds2 Offline Aishell ASR0 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr0/ds2.model.tar.gz
)
| Aishell Dataset | Char-based | 306 MB | 2 Conv + 3 bidirectional GRU layers| 0.064 |-| 151 h |
[
Ds2 Offline Aishell ASR0
](
../../examples/aishell/asr0
)
[
Conformer Online Aishell ASR1 Model
](
https://deepspeech.bj.bcebos.com/release2.1/aishell/s1/aishell.chunk.release.tar.gz
)
| Aishell Dataset | Char-based | 283 MB | Encoder:Conformer, Decoder:Transformer, Decoding method: Attention rescoring | 0.0594 |-| 151 h |
[
Conformer Online Aishell ASR1
](
../../examples/aishell/asr1
)
[
Conformer Offline Aishell ASR1 Model
](
https://deepspeech.bj.bcebos.com/release2.1/aishell/s1/aishell.release.tar.gz
)
| Aishell Dataset | Char-based | 284 MB | Encoder:Conformer, Decoder:Transformer, Decoding method: Attention rescoring | 0.0547 |-| 151 h |
[
Conformer Offline Aishell ASR1
](
../../examples/aishell/asr1
)
[
Transformer Aishell ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr1/transformer.model.tar.gz
)
| Aishell Dataset | Char-based | 128 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: Attention rescoring | 0.0523 || 151 h |
[
Transformer Aishell ASR1
](
../../examples/aishell/asr1
)
[
Conformer Librispeech ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr1/conformer.model.tar.gz
)
| Librispeech Dataset | subword-based | 191 MB | Encoder:Conformer, Decoder:Transformer, Decoding method: Attention rescoring |-| 0.0337 | 960 h |
[
Conformer Librispeech ASR1
](
../../example/librispeech/asr1
)
[
Transformer Librispeech ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr1/transformer.model.tar.gz
)
| Librispeech Dataset | subword-based | 131 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: Attention rescoring |-| 0.0381 | 960 h |
[
Transformer Librispeech ASR1
](
../../example/librispeech/asr1
)
[
Transformer Librispeech ASR2 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr2/transformer.model.tar.gz
)
| Librispeech Dataset | subword-based | 131 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: JoinCTC w/ LM |-| 0.0240 | 960 h |
[
Transformer Librispeech ASR2
](
../../example/librispeech/asr2
)
[
Ds2 Online Aishell ASR0 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr0/asr0_deepspeech2_online_aishell_ckpt_0.1.1.model.tar.gz
)
| Aishell Dataset | Char-based | 345 MB | 2 Conv + 5 LSTM layers with only forward direction | 0.080 |-| 151 h |
[
D2 Online Aishell ASR0
](
../../examples/aishell/asr0
)
[
Ds2 Offline Aishell ASR0 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr0/asr0_deepspeech2_aishell_ckpt_0.1.1.model.tar.gz
)
| Aishell Dataset | Char-based | 306 MB | 2 Conv + 3 bidirectional GRU layers| 0.064 |-| 151 h |
[
Ds2 Offline Aishell ASR0
](
../../examples/aishell/asr0
)
[
Conformer Offline Aishell ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr1/asr1_conformer_aishell_ckpt_0.1.1.model.tar.gz
)
| Aishell Dataset | Char-based | 284 MB | Encoder:Conformer, Decoder:Transformer, Decoding method: Attention rescoring | 0.056 |-| 151 h |
[
Conformer Offline Aishell ASR1
](
../../examples/aishell/asr1
)
[
Transformer Aishell ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/aishell/asr1/asr1_transformer_aishell_ckpt_0.1.1.model.tar.gz
)
| Aishell Dataset | Char-based | 128 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: Attention rescoring | 0.0523 || 151 h |
[
Transformer Aishell ASR1
](
../../examples/aishell/asr1
)
[
Conformer Librispeech ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr1/asr1_conformer_librispeech_ckpt_0.1.1.model.tar.gz
)
| Librispeech Dataset | subword-based | 191 MB | Encoder:Conformer, Decoder:Transformer, Decoding method: Attention rescoring |-| 0.0337 | 960 h |
[
Conformer Librispeech ASR1
](
../../example/librispeech/asr1
)
[
Transformer Librispeech ASR1 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr1/asr1_transformer_librispeech_ckpt_0.1.1.model.tar.gz
)
| Librispeech Dataset | subword-based | 131 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: Attention rescoring |-| 0.0381 | 960 h |
[
Transformer Librispeech ASR1
](
../../example/librispeech/asr1
)
[
Transformer Librispeech ASR2 Model
](
https://paddlespeech.bj.bcebos.com/s2t/librispeech/asr2/asr2_transformer_librispeech_ckpt_0.1.1.model.tar.gz
)
| Librispeech Dataset | subword-based | 131 MB | Encoder:Transformer, Decoder:Transformer, Decoding method: JoinCTC w/ LM |-| 0.0240 | 960 h |
[
Transformer Librispeech ASR2
](
../../example/librispeech/asr2
)
### Language Model based on NGram
Language Model | Training Data | Token-based | Size | Descriptions
...
...
@@ -25,7 +24,7 @@ Language Model | Training Data | Token-based | Size | Descriptions
| Model | Training Data | Token-based | Size | Descriptions | BLEU | Example Link |
| :-----: | :-----: | :-----: | :-----: | :-----: | :-----: | :-----: |
|
[
Transformer FAT-ST MTL En-Zh
](
https://paddlespeech.bj.bcebos.com/s2t/ted_en_zh/st1/
fat_st_ted-en-zh
.tar.gz
)
| Ted-En-Zh| Spm| | Encoder:Transformer, Decoder:Transformer,
<br
/>
Decoding method: Attention | 20.80 |
[
Transformer Ted-En-Zh ST1
](
https://github.com/PaddlePaddle/PaddleSpeech/blob/develop/examples/ted_en_zh/st1
)
|
|
[
Transformer FAT-ST MTL En-Zh
](
https://paddlespeech.bj.bcebos.com/s2t/ted_en_zh/st1/
st1_transformer_mtl_noam_ted-en-zh_ckpt_0.1.1.model
.tar.gz
)
| Ted-En-Zh| Spm| | Encoder:Transformer, Decoder:Transformer,
<br
/>
Decoding method: Attention | 20.80 |
[
Transformer Ted-En-Zh ST1
](
https://github.com/PaddlePaddle/PaddleSpeech/blob/develop/examples/ted_en_zh/st1
)
|
## Text-to-Speech Models
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录