paddle_nccl.md 4.0 KB
Newer Older
D
dzhwinter 已提交
1
# Design Doc: NCCL support in Paddle Fluid
D
dongzhihong 已提交
2 3 4

## Abstract

D
dzhwinter 已提交
5
This Design Doc refers to the NCCL feature in  paddle.  We propose an approach to support NCCL library both on a single machine and multiple machines. We wrapper the NCCL primitives `Broadcast`, `Allreduce`, `Reduce` as operators to utilize Multi-GPU powers in one script.
D
dongzhihong 已提交
6 7 8 9


## Motivation

D
dzhwinter 已提交
10
NCCL is a NVIDIA library support Multi-GPU communicating and optimized for NVIDIA GPUs, it provides routines such as all-gather, all-reduce, broadcast, reduce, reduce-scatter, that can achieve high bandwidth over PCIe and NVLink high-speed interconnect. [NCCL](https://developer.nvidia.com/nccl). With NCCL library, we can easily accelerate the training in parallel. 
D
dongzhihong 已提交
11

D
dzhwinter 已提交
12 13 14 15
- Pros
1. easily plug-in with [NCCL2](https://developer.nvidia.com/nccl) library.
1. high performance in NVIDIA GPUs.
1. MPI like primitives, which have low learning cost for users.
D
dongzhihong 已提交
16

D
dzhwinter 已提交
17 18 19
- Cons
1. Only design for NVIDIA GPUs, not a general multi-device solution.
1. Although NCCL1 is opensourced under BSD license, but NCCL2 is not opensourced anymore.
D
dongzhihong 已提交
20

D
dzhwinter 已提交
21
At the beginning of training, the framework needs to distribute the same parameters to every GPU, and merge the gradients at any time user interests.
D
dongzhihong 已提交
22

D
dzhwinter 已提交
23
As a result, during training, we need the operations of peer to peer copy between different GPUs, aggregating gradients/parameters from GPUs, and broadcasting parameters to GPUs. Every GPU only need to run the operator with correct place information.
D
dongzhihong 已提交
24

D
dzhwinter 已提交
25
Besides, it needs interfaces to synchronize model update with each different GPU Cards. 
D
dongzhihong 已提交
26 27 28

## Implementation

D
dzhwinter 已提交
29
As mentioned above, we wrap the NCCL routines as several kinds of operators. Need to note that NCCL need to create Communicator between gpu at the beginning, so there is a NCCLInit operator created.
D
dongzhihong 已提交
30 31 32

### Graph Converter

D
dongzhihong 已提交
33
To be compatible with [parameter server design doc](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/design/ops/dist_train.md), the graph converter converts the user defined operation graph into sub-graphs to be executed on different devices.
D
dongzhihong 已提交
34

D
dzhwinter 已提交
35
1. The user-defined model will be a single device program
D
dongzhihong 已提交
36

D
dzhwinter 已提交
37
2. Broadcast/Reduce operators between GPUs will be inserted into the program, even for the multi-node, may insert the `Send`, `Recv` operator.
D
dongzhihong 已提交
38

D
dongzhihong 已提交
39
   *Broadcast, AllReduce in a single machine. And Broadcast, AllReduce, [Send, Recv](https://github.com/PaddlePaddle/Paddle/blob/develop/doc/design/ops/dist_train.md#graph-converter) in multiple machines*
D
dongzhihong 已提交
40 41 42 43 44 45 46 47 48 49

   <img src="images/multigpu_before_convert.png" width="300"/>

After convert, the graph as shows

<img src="images/multigpu_allreduce.png" width="1000"/>

Operators are added to the sub-graphs. Every GPU assigned a role of `rank0`, `rank1` etc. 

- **Broadcast**. Broadcast operator distribute initialized parameter to all the GPUs from the GPU who owns it. e.g. from`rank0` GPU.
D
dzhwinter 已提交
50
- **AllReduce**. AllReduce operator synchronizes parameters/gradients between GPUs. AllReduce implemented in the Ring-Based  communicating method, avoid of the bottle neck in a single GPU.
D
dongzhihong 已提交
51

D
dzhwinter 已提交
52
Need to notice that AllReduce operator force GPUs synchronized at that point. The whole training process in asynchronous or synchronous mode depends on the AllReduce point in the graph.
D
dongzhihong 已提交
53

D
dongzhihong 已提交
54
As it shown in the picture, when each GPU compute the gradient of `W`, followed with a `AllReduce` operator, accumulate the `dW` to full batch of data, then run the optimize process individually and apply the gradient to its `W`.
D
dongzhihong 已提交
55

D
dzhwinter 已提交
56 57 58 59 60
- **AllReduce2**
If we use the NCCL2 AllReduce primitive, every GPU optimized full batch of data, wasted (n-1) GPU compute resources. In addition, AllReduce will only utilize the communicate resource during synchronization, then update the gradient will be a seperated phase. In fact, we can amortize the update gradient time cost into the communicating phase.
- Every parameter has its root card. That card will call **Reduce** operator and collect the gradients from GPUs.
- The whole model's parameter will be hashed to different root card, ensure the load balance between GPUs.
Then we have another version AllReduce operator. Other part keep the same with before.