README.md 24.5 KB
Newer Older
1
# DeepSpeech2 on PaddlePaddle
2

L
lfchener 已提交
3 4
[中文版](README_cn.md)

L
lfchener 已提交
5
*DeepSpeech2 on PaddlePaddle* is an open-source implementation of end-to-end Automatic Speech Recognition (ASR) engine, based on [Baidu's Deep Speech 2 paper](http://proceedings.mlr.press/v48/amodei16.pdf), with [PaddlePaddle](https://github.com/PaddlePaddle/Paddle) platform. Our vision is to empower both industrial application and academic research on speech recognition, via an easy-to-use, efficient and scalable implementation, including training, inference & testing module, and demo deployment. Besides, several pre-trained models for both English and Mandarin are also released.
6 7 8 9 10 11

## Table of Contents
- [Installation](#installation)
- [Getting Started](#getting-started)
- [Data Preparation](#data-preparation)
- [Training a Model](#training-a-model)
12
- [Data Augmentation Pipeline](#data-augmentation-pipeline)
13
- [Inference and Evaluation](#inference-and-evaluation)
14
- [Running in Docker Container](#running-in-docker-container)
15
- [Hyper-parameters Tuning](#hyper-parameters-tuning)
16
- [Training for Mandarin Language](#training-for-mandarin-language)
17
- [Trying Live Demo with Your Own Voice](#trying-live-demo-with-your-own-voice)
18
- [Released Models](#released-models)
19
- [Experiments and Benchmarks](#experiments-and-benchmarks)
20 21
- [Questions and Help](#questions-and-help)

22

23

24
## Installation
25

L
lfchener 已提交
26 27 28
To avoid the trouble of environment setup, [running in Docker container](#running-in-docker-container) is highly recommended. Otherwise follow the guidelines below to install the dependencies manually.

### Prerequisites
H
Hui Zhang 已提交
29
- Python >= 3.5
L
lfchener 已提交
30
- PaddlePaddle 1.8.0 or later (please refer to the [Installation Guide](https://www.paddlepaddle.org.cn/documentation/docs/en/beginners_guide/index_en.html))
L
lfchener 已提交
31 32 33 34 35

### Setup
- Make sure these libraries or tools installed: `pkg-config`, `flac`, `ogg`, `vorbis`, `boost` and `swig`, e.g. installing them via `apt-get`:

```bash
H
Hui Zhang 已提交
36
sudo apt-get install -y pkg-config libflac-dev libogg-dev libvorbis-dev libboost-dev swig python3-dev
L
lfchener 已提交
37 38 39 40 41
```

or, installing them via `yum`:

```bash
H
Hui Zhang 已提交
42
sudo yum install pkgconfig libogg-devel libvorbis-devel boost-devel python3-devel
L
lfchener 已提交
43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58
wget https://ftp.osuosl.org/pub/xiph/releases/flac/flac-1.3.1.tar.xz
xz -d flac-1.3.1.tar.xz
tar -xvf flac-1.3.1.tar
cd flac-1.3.1
./configure
make
make install
```

- Run the setup script for the remaining dependencies

```bash
git clone https://github.com/PaddlePaddle/DeepSpeech.git
cd DeepSpeech
sh setup.sh
```
59

H
Hui Zhang 已提交
60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92
### Running in Docker Container

Docker is an open source tool to build, ship, and run distributed applications in an isolated environment. A Docker image for this project has been provided in [hub.docker.com](https://hub.docker.com) with all the dependencies installed, including the pre-built PaddlePaddle, CTC decoders, and other necessary Python and third-party packages. This Docker image requires the support of NVIDIA GPU, so please make sure its availiability and the [nvidia-docker](https://github.com/NVIDIA/nvidia-docker) has been installed.

Take several steps to launch the Docker image:

- Download the Docker image

```bash
nvidia-docker pull hub.baidubce.com/paddlepaddle/deep_speech_fluid:latest-gpu
```

- Clone this repository

```
git clone https://github.com/PaddlePaddle/DeepSpeech.git
```

- Run the Docker image

```bash
sudo nvidia-docker run -it -v $(pwd)/DeepSpeech:/DeepSpeech hub.baidubce.com/paddlepaddle/deep_speech_fluid:latest-gpu /bin/bash
```
Now go back and start from the [Getting Started](#getting-started) section, you can execute training, inference and hyper-parameters tuning similarly in the Docker container.


- Install PaddlePaddle

For example, for CUDA 10.1, CuDNN7.5:
```bash
python3 -m pip install paddlepaddle-gpu==1.8.0.post107
```

93
## Getting Started
94

95
Several shell scripts provided in `./examples` will help us to quickly give it a try, for most major modules, including data preparation, model training, case inference and model evaluation, with a few public dataset (e.g. [LibriSpeech](http://www.openslr.org/12/), [Aishell](http://www.openslr.org/33)). Reading these examples will also help you to understand how to make it work with your own data.
96

L
lfchener 已提交
97
Some of the scripts in `./examples` are configured with 8 GPUs. If you don't have 8 GPUs available, please modify `CUDA_VISIBLE_DEVICES`. If you don't have any GPU available, please set `--use_gpu` to False to use CPUs instead. Besides, if out-of-memory problem occurs, just reduce `--batch_size` to fit.
98 99 100 101 102

Let's take a tiny sampled subset of [LibriSpeech dataset](http://www.openslr.org/12/) for instance.

- Go to directory

103
    ```bash
104
    cd examples/tiny
105 106
    ```

107
    Notice that this is only a toy example with a tiny sampled subset of LibriSpeech. If you would like to try with the complete dataset (would take several days for training), please go to `examples/librispeech` instead.
108
- Prepare the data
109

110
    ```bash
111
    sh run_data.sh
112 113
    ```

L
lfchener 已提交
114
    `run_data.sh` will download dataset, generate manifests, collect normalizer's statistics and build vocabulary. Once the data preparation is done, you will find the data (only part of LibriSpeech) downloaded in `./dataset/librispeech` and the corresponding manifest files generated in `./data/tiny` as well as a mean stddev file and a vocabulary file. It has to be run for the very first time you run this dataset and is reusable for all further experiments.
115 116
- Train your own ASR model

117
    ```bash
118 119 120
    sh run_train.sh
    ```

121
    `run_train.sh` will start a training job, with training logs printed to stdout and model checkpoint of every pass/epoch saved to `./checkpoints/tiny`. These checkpoints could be used for training resuming, inference, evaluation and deployment.
122 123
- Case inference with an existing model

124
    ```bash
125 126 127
    sh run_infer.sh
    ```

128
    `run_infer.sh` will show us some speech-to-text decoding results for several (default: 10) samples with the trained model. The performance might not be good now as the current model is only trained with a toy subset of LibriSpeech. To see the results with a better model, you can download a well-trained (trained for several days, with the complete LibriSpeech) model and do the inference:
129

130
    ```bash
131
    sh run_infer_golden.sh
132 133 134
    ```
- Evaluate an existing model

135
    ```bash
136 137 138
    sh run_test.sh
    ```

139
    `run_test.sh` will evaluate the model with Word Error Rate (or Character Error Rate) measurement. Similarly, you can also download a well-trained model and test its performance:
140

141
    ```bash
142
    sh run_test_golden.sh
143 144
    ```

X
Xinghai Sun 已提交
145
More detailed information are provided in the following sections. Wish you a happy journey with the *DeepSpeech2 on PaddlePaddle* ASR engine!
146

147

148
## Data Preparation
149

X
Xinghai Sun 已提交
150
### Generate Manifest
151

152
*DeepSpeech2 on PaddlePaddle* accepts a textual **manifest** file as its data set interface. A manifest file summarizes a set of speech data, with each line containing some meta data (e.g. filepath, transcription, duration) of one audio clip, in [JSON](http://www.json.org/) format, such as:
153

154
```
155 156
{"audio_filepath": "/home/work/.cache/paddle/Libri/134686/1089-134686-0001.flac", "duration": 3.275, "text": "stuff it into you his belly counselled him"}
{"audio_filepath": "/home/work/.cache/paddle/Libri/134686/1089-134686-0007.flac", "duration": 4.275, "text": "a cold lucid indifference reigned in his soul"}
157
```
158

159
To use your custom data, you only need to generate such manifest files to summarize the dataset. Given such summarized manifests, training, inference and all other modules can be aware of where to access the audio files, as well as their meta data including the transcription labels.
160

161
For how to generate such manifest files, please refer to `data/librispeech/librispeech.py`, which will download data and generate manifest files for LibriSpeech dataset.
X
Xinghai Sun 已提交
162

X
Xinghai Sun 已提交
163
### Compute Mean & Stddev for Normalizer
X
Xinghai Sun 已提交
164

165
To perform z-score normalization (zero-mean, unit stddev) upon audio features, we have to estimate in advance the mean and standard deviation of the features, with some training samples:
X
Xinghai Sun 已提交
166

167
```bash
H
Hui Zhang 已提交
168
python3 tools/compute_mean_std.py \
169 170
--num_samples 2000 \
--specgram_type linear \
A
AshishKarel 已提交
171
--manifest_path data/librispeech/manifest.train \
172
--output_path data/librispeech/mean_std.npz
X
Xinghai Sun 已提交
173 174
```

L
lfchener 已提交
175
It will compute the mean and standard deviatio of power spectrum feature with 2000 random sampled audio clips listed in `data/librispeech/manifest.train` and save the results to `data/librispeech/mean_std.npz` for further usage.
X
Xinghai Sun 已提交
176

177

X
Xinghai Sun 已提交
178
### Build Vocabulary
X
Xinghai Sun 已提交
179

X
Xinghai Sun 已提交
180
A vocabulary of possible characters is required to convert the transcription into a list of token indices for training, and in decoding, to convert from a list of indices back to text again. Such a character-based vocabulary can be built with `tools/build_vocab.py`.
Y
Yibing Liu 已提交
181

182
```bash
H
Hui Zhang 已提交
183
python3 tools/build_vocab.py \
184 185 186
--count_threshold 0 \
--vocab_path data/librispeech/eng_vocab.txt \
--manifest_paths data/librispeech/manifest.train
Y
Yibing Liu 已提交
187
```
188

189
It will write a vocabuary file `data/librispeeech/eng_vocab.txt` with all transcription text in `data/librispeech/manifest.train`, without vocabulary truncation (`--count_threshold 0`).
190

X
Xinghai Sun 已提交
191
### More Help
192

193
For more help on arguments:
194

195
```bash
H
Hui Zhang 已提交
196 197 198
python3 data/librispeech/librispeech.py --help
python3 tools/compute_mean_std.py --help
python3 tools/build_vocab.py --help
199 200
```

201
## Training a model
202

203
`train.py` is the main caller of the training module. Examples of usage are shown below.
204

205
- Start training from scratch with 8 GPUs:
206

207
    ```
H
Hui Zhang 已提交
208
    CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python3 train.py
209
    ```
210

L
lfchener 已提交
211
- Start training from scratch with CPUs:
212 213

    ```
H
Hui Zhang 已提交
214
    python3 train.py --use_gpu False
215
    ```
216
- Resume training from a checkpoint:
217 218

    ```
X
Xinghai Sun 已提交
219
    CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \
H
Hui Zhang 已提交
220
    python3 train.py \
221
    --init_from_pretrained_model CHECKPOINT_PATH_TO_RESUME_FROM
222
    ```
223

224
For more help on arguments:
225

226
```bash
H
Hui Zhang 已提交
227
python3 train.py --help
228
```
229
or refer to `example/librispeech/run_train.sh`.
230

L
lfchener 已提交
231

232
## Data Augmentation Pipeline
Y
Yibing Liu 已提交
233

234
Data augmentation has often been a highly effective technique to boost the deep learning performance. We augment our speech data by synthesizing new audios with small random perturbation (label-invariant transformation) added upon raw audios. You don't have to do the syntheses on your own, as it is already embedded into the data provider and is done on the fly, randomly for each epoch during training.
235

236
Six optional augmentation components are provided to be selected, configured and inserted into the processing pipeline.
237 238 239 240

  - Volume Perturbation
  - Speed Perturbation
  - Shifting Perturbation
X
Xinghai Sun 已提交
241
  - Online Bayesian normalization
242 243 244
  - Noise Perturbation (need background noise audio files)
  - Impulse Response (need impulse audio files)

H
Hu Weiwei 已提交
245
In order to inform the trainer of what augmentation components are needed and what their processing orders are, it is required to prepare in advance an *augmentation configuration file* in [JSON](http://www.json.org/) format. For example:
Y
Yibing Liu 已提交
246 247

```
248 249 250 251 252 253 254 255 256 257 258 259
[{
    "type": "speed",
    "params": {"min_speed_rate": 0.95,
               "max_speed_rate": 1.05},
    "prob": 0.6
},
{
    "type": "shift",
    "params": {"min_shift_ms": -5,
               "max_shift_ms": 5},
    "prob": 0.8
}]
Y
Yibing Liu 已提交
260 261
```

262
When the `--augment_conf_file` argument of `trainer.py` is set to the path of the above example configuration file, every audio clip in every epoch will be processed: with 60% of chance, it will first be speed perturbed with a uniformly random sampled speed-rate between 0.95 and 1.05, and then with 80% of chance it will be shifted in time with a random sampled offset between -5 ms and 5 ms. Finally this newly synthesized audio clip will be feed into the feature extractor for further training.
263

264
For other configuration examples, please refer to `conf/augmenatation.config.example`.
Y
Yibing Liu 已提交
265

266
Be careful when utilizing the data augmentation technique, as improper augmentation will do harm to the training, due to the enlarged train-test gap.
267

268
## Inference and Evaluation
269

270
### Prepare Language Model
Y
Yibing Liu 已提交
271

272
A language model is required to improve the decoder's performance. We have prepared two language models (with lossy compression) for users to download and try. One is for English and the other is for Mandarin. Users can simply run this to download the preprared language models:
X
Xinghai Sun 已提交
273

274
```bash
X
Xinghai Sun 已提交
275
cd models/lm
L
lfchener 已提交
276 277
bash download_lm_en.sh
bash download_lm_ch.sh
X
Xinghai Sun 已提交
278
```
279

Y
yangyaming 已提交
280
If you wish to train your own better language model, please refer to [KenLM](https://github.com/kpu/kenlm) for tutorials. Here we provide some tips to show how we preparing our English and Mandarin language models. You can take it as a reference when you train your own.
Y
yangyaming 已提交
281 282 283

#### English LM

H
Hu Weiwei 已提交
284
The English corpus is from the [Common Crawl Repository](http://commoncrawl.org) and you can download it from [statmt](http://data.statmt.org/ngrams/deduped_en). We use part en.00 to train our English language model. There are some preprocessing steps before training:
Y
yangyaming 已提交
285

L
lfchener 已提交
286
  * Characters not in \['A-Za-z0-9\s'\] (\s represents whitespace characters) are removed and Arabic numbers are converted to English numbers like 1000 to one thousand.
Y
yangyaming 已提交
287 288
  * Repeated whitespace characters are squeezed to one and the beginning whitespace characters are removed. Notice that all transcriptions are lowercase, so all characters are converted to lowercase.
  * Top 400,000 most frequent words are selected to build the vocabulary and the rest are replaced with 'UNKNOWNWORD'.
Y
yangyaming 已提交
289

Y
yangyaming 已提交
290
Now the preprocessing is done and we get a clean corpus to train the language model. Our released language model are trained with agruments '-o 5 --prune 0 1 1 1 1'. '-o 5' means the max order of language model is 5. '--prune 0 1 1 1 1' represents count thresholds for each order and more specifically it will prune singletons for orders two and higher. To save disk storage we convert the arpa file to 'trie' binary file with arguments '-a 22 -q 8 -b 8'. '-a' represents the maximum number of leading bits of pointers in 'trie' to chop. '-q -b' are quantization parameters for probability and backoff.
Y
yangyaming 已提交
291

Y
yangyaming 已提交
292 293
#### Mandarin LM

Y
yangyaming 已提交
294
Different from the English language model, Mandarin language model is character-based where each token is a Chinese character. We use internal corpus to train the released Mandarin language models. The corpus contain billions of tokens. The preprocessing has tiny difference from English language model and main steps include:
Y
yangyaming 已提交
295 296

  * The beginning and trailing whitespace characters are removed.
Y
yangyaming 已提交
297 298
  * English punctuations and Chinese punctuations are removed.
  * A whitespace character between two tokens is inserted.
Y
yangyaming 已提交
299

Y
yangyaming 已提交
300
Please notice that the released language models only contain Chinese simplified characters. After preprocessing done we can begin to train the language model. The key training arguments for small LM is '-o 5 --prune 0 1 2 4 4' and '-o 5' for large LM. Please refer above section for the meaning of each argument. We also convert the arpa file to binary file using default settings.
301

302
### Speech-to-text Inference
303

304
An inference module caller `infer.py` is provided to infer, decode and visualize speech-to-text results for several given audio clips. It might help to have an intuitive and qualitative evaluation of the ASR model's performance.
305 306 307

- Inference with GPU:

308
    ```bash
H
Hui Zhang 已提交
309
    CUDA_VISIBLE_DEVICES=0 python3 infer.py
310
    ```
Y
Yibing Liu 已提交
311

X
Xinghai Sun 已提交
312
- Inference with CPUs:
313

314
    ```bash
H
Hui Zhang 已提交
315
    python3 infer.py --use_gpu False
316 317
    ```

X
Xinghai Sun 已提交
318
We provide two types of CTC decoders: *CTC greedy decoder* and *CTC beam search decoder*. The *CTC greedy decoder* is an implementation of the simple best-path decoding algorithm, selecting at each timestep the most likely token, thus being greedy and locally optimal. The [*CTC beam search decoder*](https://arxiv.org/abs/1408.2873) otherwise utilizes a heuristic breadth-first graph search for reaching a near global optimality; it also requires a pre-trained KenLM language model for better scoring and ranking. The decoder type can be set with argument `--decoding_method`.
319 320

For more help on arguments:
321 322

```
H
Hui Zhang 已提交
323
python3 infer.py --help
324
```
325
or refer to `example/librispeech/run_infer.sh`.
Y
Yibing Liu 已提交
326

327
### Evaluate a Model
Y
Yibing Liu 已提交
328

329
To evaluate a model's performance quantitatively, please run:
330

X
Xinghai Sun 已提交
331
- Evaluation with GPUs:
332

333
    ```bash
H
Hui Zhang 已提交
334
    CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 python3 test.py
335 336
    ```

X
Xinghai Sun 已提交
337
- Evaluation with CPUs:
Y
Yibing Liu 已提交
338

339
    ```bash
H
Hui Zhang 已提交
340
    python3 test.py --use_gpu False
341 342
    ```

343
The error rate (default: word error rate; can be set with `--error_rate_type`) will be printed.
344 345

For more help on arguments:
Y
Yibing Liu 已提交
346

347
```bash
H
Hui Zhang 已提交
348
python3 test.py --help
Y
Yibing Liu 已提交
349
```
350
or refer to `example/librispeech/run_test.sh`.
Y
Yibing Liu 已提交
351

352
## Hyper-parameters Tuning
Y
Yibing Liu 已提交
353

354
The hyper-parameters $\alpha$ (language model weight) and $\beta$ (word insertion weight) for the [*CTC beam search decoder*](https://arxiv.org/abs/1408.2873) often have a significant impact on the decoder's performance. It would be better to re-tune them on the validation set when the acoustic model is renewed.
Y
Yibing Liu 已提交
355

356
`tools/tune.py` performs a 2-D grid search over the hyper-parameter $\alpha$ and $\beta$. You must provide the range of $\alpha$ and $\beta$, as well as the number of their attempts.
Y
Yibing Liu 已提交
357

358
- Tuning with GPU:
Y
Yibing Liu 已提交
359

360
    ```bash
X
Xinghai Sun 已提交
361
    CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7 \
H
Hui Zhang 已提交
362
    python3 tools/tune.py \
363 364 365 366 367 368
    --alpha_from 1.0 \
    --alpha_to 3.2 \
    --num_alphas 45 \
    --beta_from 0.1 \
    --beta_to 0.45 \
    --num_betas 8
369
    ```
Y
Yibing Liu 已提交
370

371
- Tuning with CPU:
Y
Yibing Liu 已提交
372

373
    ```bash
H
Hui Zhang 已提交
374
    python3 tools/tune.py --use_gpu False
375
    ```
Y
yangyaming 已提交
376
 The grid search will print the WER (word error rate) or CER (character error rate) at each point in the hyper-parameters space, and draw the error surface optionally. A proper hyper-parameters range should include the global minima of the error surface for WER/CER, as illustrated in the following figure.
377

378
<p align="center">
Y
Yibing Liu 已提交
379
<img src="docs/images/tuning_error_surface.png" width=550>
380 381 382
<br/>An example error surface for tuning on the dev-clean set of LibriSpeech
</p>

Y
Yibing Liu 已提交
383
Usually, as the figure shows, the variation of language model weight ($\alpha$) significantly affect the performance of CTC beam search decoder. And a better procedure is to first tune on serveral data batches (the number can be specified) to find out the proper range of hyper-parameters, then change to the whole validation set to carray out an accurate tuning.
384 385

After tuning, you can reset $\alpha$ and $\beta$ in the inference and evaluation modules to see if they really help improve the ASR performance. For more help
Y
Yibing Liu 已提交
386

387
```bash
H
Hui Zhang 已提交
388
python3 tune.py --help
Y
Yibing Liu 已提交
389
```
390
or refer to `example/librispeech/run_tune.sh`.
Y
Yibing Liu 已提交
391

392 393
## Training for Mandarin Language

Y
yangyaming 已提交
394
The key steps of training for Mandarin language are same to that of English language and we have also provided an example for Mandarin training with Aishell in ```examples/aishell```. As mentioned above, please execute ```sh run_data.sh```, ```sh run_train.sh```, ```sh run_test.sh``` and ```sh run_infer.sh``` to do data preparation, training, testing and inference correspondingly. We have also prepared a pre-trained model (downloaded by ./models/aishell/download_model.sh) for users to try with ```sh run_infer_golden.sh``` and ```sh run_test_golden.sh```. Notice that, different from English LM, the Mandarin LM is character-based and please run ```tools/tune.py``` to find an optimal setting.
X
Xinghai Sun 已提交
395

396
## Trying Live Demo with Your Own Voice
397

L
lfchener 已提交
398
Until now, an ASR model is trained and tested qualitatively (`infer.py`) and quantitatively (`test.py`) with existing audio files. But it is not yet tested with your own speech. `deploy/demo_english_server.py` and `deploy/demo_client.py` helps quickly build up a real-time demo ASR engine with the trained model, enabling you to test and play around with the demo, with your own voice.
X
Xinghai Sun 已提交
399

400
To start the demo's server, please run this in one console:
X
Xinghai Sun 已提交
401

402
```bash
X
Xinghai Sun 已提交
403
CUDA_VISIBLE_DEVICES=0 \
H
Hui Zhang 已提交
404
python3 deploy/demo_server.py \
X
Xinghai Sun 已提交
405 406 407 408
--host_ip localhost \
--host_port 8086
```

409
For the machine (might not be the same machine) to run the demo's client, please do the following installation before moving on.
410 411 412

For example, on MAC OS X:

413
```bash
414 415
brew install portaudio
pip install pyaudio
L
lfchener 已提交
416
pip install keyboard
417
```
X
Xinghai Sun 已提交
418

419
Then to start the client, please run this in another console:
420

421
```bash
X
Xinghai Sun 已提交
422
CUDA_VISIBLE_DEVICES=0 \
H
Hui Zhang 已提交
423
python3 -u deploy/demo_client.py \
X
Xinghai Sun 已提交
424
--host_ip 'localhost' \
X
Xinghai Sun 已提交
425
--host_port 8086
426
```
X
Xinghai Sun 已提交
427

428
Now, in the client console, press the `whitespace` key, hold, and start speaking. Until finishing your utterance, release the key to let the speech-to-text results shown in the console. To quit the client, just press `ESC` key.
X
Xinghai Sun 已提交
429

430
Notice that `deploy/demo_client.py` must be run on a machine with a microphone device, while `deploy/demo_server.py` could be run on one without any audio recording hardware, e.g. any remote server machine. Just be careful to set the `host_ip` and `host_port` argument with the actual accessible IP address and port, if the server and client are running with two separate machines. Nothing should be done if they are running on one single machine.
X
Xinghai Sun 已提交
431

L
lfchener 已提交
432
Please also refer to `examples/deploy_demo/run_english_demo_server.sh`, which will first download a pre-trained English model (trained with 3000 hours of internal speech data) and then start the demo server with the model. With running `examples/mandarin/run_demo_client.sh`, you can speak English to test it. If you would like to try some other models, just update `--model_path` argument in the script.  
X
Xinghai Sun 已提交
433 434

For more help on arguments:
435

436
```bash
H
Hui Zhang 已提交
437 438
python3 deploy/demo_server.py --help
python3 deploy/demo_client.py --help
439
```
440

441 442 443 444
## Released Models

#### Speech Model Released

X
Xinghai Sun 已提交
445
Language  | Model Name | Training Data | Hours of Speech
446
:-----------: | :------------: | :----------: |  -------:
L
lfchener 已提交
447 448 449 450
English  | [LibriSpeech Model](https://deepspeech.bj.bcebos.com/eng_models/librispeech_model_fluid.tar.gz) | [LibriSpeech Dataset](http://www.openslr.org/12/) | 960 h
English  | [BaiduEN8k Model](https://deepspeech.bj.bcebos.com/demo_models/baidu_en8k_model_fluid.tar.gz) | Baidu Internal English Dataset | 8628 h
Mandarin | [Aishell Model](https://deepspeech.bj.bcebos.com/mandarin_models/aishell_model_fluid.tar.gz) | [Aishell Dataset](http://www.openslr.org/33/) | 151 h
Mandarin | [BaiduCN1.2k Model](https://deepspeech.bj.bcebos.com/demo_models/baidu_cn1.2k_model_fluid.tar.gz) | Baidu Internal Mandarin Dataset | 1204 h
451 452 453

#### Language Model Released

Y
yangyaming 已提交
454 455
Language Model | Training Data | Token-based | Size | Descriptions
:-------------:| :------------:| :-----: | -----: | :-----------------
Y
Yibing Liu 已提交
456 457 458
[English LM](https://deepspeech.bj.bcebos.com/en_lm/common_crawl_00.prune01111.trie.klm) |  [CommonCrawl(en.00)](http://web-language-models.s3-website-us-east-1.amazonaws.com/ngrams/en/deduped/en.00.deduped.xz) | Word-based | 8.3 GB | Pruned with 0 1 1 1 1; <br/> About 1.85 billion n-grams; <br/> 'trie'  binary with '-a 22 -q 8 -b 8'
[Mandarin LM Small](https://deepspeech.bj.bcebos.com/zh_lm/zh_giga.no_cna_cmn.prune01244.klm) | Baidu Internal Corpus | Char-based | 2.8 GB | Pruned with 0 1 2 4 4; <br/> About 0.13 billion n-grams; <br/> 'probing' binary with default settings
[Mandarin LM Large](https://deepspeech.bj.bcebos.com/zh_lm/zhidao_giga.klm) | Baidu Internal Corpus | Char-based | 70.4 GB | No Pruning; <br/> About 3.7 billion n-grams; <br/> 'probing' binary with default settings
459

460
## Experiments and Benchmarks
461

X
Xinghai Sun 已提交
462
#### Benchmark Results for English Models (Word Error Rate)
X
Xinghai Sun 已提交
463

X
Xinghai Sun 已提交
464
Test Set                | LibriSpeech Model | BaiduEN8K Model
X
Xinghai Sun 已提交
465
:---------------------  | ---------------:  | -------------------:
466 467 468 469 470 471 472
LibriSpeech Test-Clean  |   6.85            |   5.41
LibriSpeech Test-Other  |   21.18           |   13.85
VoxForge American-Canadian | 12.12          |   7.13
VoxForge Commonwealth   |   19.82           |   14.93
VoxForge European       |   30.15           |   18.64
VoxForge Indian         |   53.73           |   25.51
Baidu Internal Testset  |   40.75           |   8.48
473

474
For reproducing benchmark results on VoxForge data, we provide a script to download data and generate VoxForge dialect manifest files. Please go to ```data/voxforge``` and execute ```sh run_data.sh``` to get VoxForge dialect manifest files. Notice that VoxForge data may keep updating and the generated manifest files may have difference from those we evaluated on.
475

X
Xinghai Sun 已提交
476
#### Benchmark Results for Mandarin Model (Character Error Rate)
477

X
Xinghai Sun 已提交
478 479 480
Test Set                |  BaiduCN1.2k Model
:---------------------  |  -------------------:
Baidu Internal Testset  |   12.64
481

482 483
#### Acceleration with Multi-GPUs

L
lfchener 已提交
484
We compare the training time with 1, 2, 4, 8 Tesla V100 GPUs (with a subset of LibriSpeech samples whose audio durations are between 6.0 and 7.0 seconds).  And it shows that a **near-linear** acceleration with multiple GPUs has been achieved. In the following figure, the time (in seconds) cost for training is printed on the blue bars.
485 486 487 488 489 490

<img src="docs/images/multi_gpu_speedup.png" width=450><br/>

| # of GPU  | Acceleration Rate |
| --------  | --------------:   |
| 1         | 1.00 X |
L
lfchener 已提交
491 492 493
| 2         | 1.98 X |
| 4         | 3.73 X |
| 8         | 6.95 X |
494

495
`tools/profile.sh` provides such a profiling tool.
X
Xinghai Sun 已提交
496

497
## Questions and Help
X
Xinghai Sun 已提交
498

Y
Yibing Liu 已提交
499
You are welcome to submit questions and bug reports in [Github Issues](https://github.com/PaddlePaddle/DeepSpeech/issues). You are also welcome to contribute to this project.