# Design Doc: LoD (Level-of-Detail) Tensor As other deep learning systems, PaddlePaddle supports training models from sequence data. Also, like other systems, PaddlePaddle represent a mini-batch of sequences as a Tensor. What is different is that PaddlePaddle doesn't require that all sequences in a mini-batch are of the same length. Thus no need for padding zeros. | | TensorFlow | PaddlePaddle | |-----------------------|------------|--------------| | RNN | Support | Support | | recursive RNN | Support | Support | | padding zeros | Must | No need | | blob data type | Tensor | LoDTensor | PaddlePaddle achieves this flexibility by passing through a new data type, *LoD Tensor*, which is a Tensor attached with segmentation index known as *LoD*, between operators. The LoD index doesn't only segments a tensor, but also recursively segments sub-sequences. This document presents the design of LoD and LoDTensor. ## The Challenge: Variable-length Sequences Most deep learning systems represent a mini-batch as a Tensor. For example, a mini-batch of 10 images, each of size 32x32, is a 10x32x32 Tensor. Another example is that each mini-batch contains N sentences, where each word is a D-dimensional one-hot vector. Suppose that all sentences have the same length L, we can represent this mini-batch by a NxLxD tensor. Both examples show that the elements of sequences are usually of the same size. In the first example, all images are 32x32, and in the second one, all words are D-dimensional vectors. It doesn't make sense to allow variable-sized images, as that would require transformations like convolution represented by variable-sized Tensors. The real challenge is that in most cases, sentences have variable lengths, and we will need an index data structure to segment the tensor into sequences. Also, sequences might consist of sub-sequences. ## A Solution: The LoD Index Let is visit this challenge from examples. ### A Mini-Batch of Sentences Let's imagine a mini-batch of 3 variable lengths sentences composed by 3, 1, and 2 words respectively. We can represent it by a (3+1+2)xD tensor plus some index information: ``` 3 1 2 ||| | || ``` where each `|` represents a D-dimensional word vector. The numbers, 3, 1, and 2, form a 1-level LoD. ### Recursive Sequences Let check another example of a 2-level LoD Tensor. Consider a mini-batch of three articles with 3, 1, and 2 sentences, and each sentence consists of words: ``` 3 1 2 3 2 4 1 2 3 ||| || |||| | || ||| ``` ### A Mini-Batch of Videos LoD Tensor generalizes to the case where elements are higher dimensional objects, like images. Suppose that a mini-batch contains videos of the same frame size 640x480. Here is a mini-batch of 3 videos with 3, 1, and 2 frames respectively. ``` 3 1 2 口口口 口 口口 ``` The underlying tensor is of size (3+1+2)x640x480, and each `口` represents a 640x480 image. ### A Mini-Batch of Images In traditional cases like a mini-batch with N fixed-sized images, the LoD Tensor representation is as ``` 1 1 1 1 1 口口口口 ... 口 ``` It doesn't loss anything to ignore the many 1's in the index and to consider this LoD Tensor a usual Tensor: ``` 口口口口 ... 口 ``` ### Model Parameters A model parameter is just a usual Tensor, which, just like the above example, is a **0-level LoD Tensor**. ## The LoD Tensor Let us revisit above example of the 2-level LoD Tensor ``` 3 1 2 3 2 4 1 2 3 ||| || |||| | || ||| ``` It is indeed a tree, where leaves are elementary sequences identified by **branches**. For example, the third sentence in above example is identified by branch <0,2>, where 0 indicates the first article with length 3, and 2 indicates the third sentence in this article with length 4. ### The LoD Index We can save the LoD index in above example ``` 3 1 2 3 2 4 1 2 3 ``` in a not-full 2D matrix: ```c++ typedef std::vector > LoD; ``` where - `LoD.size()` is the number of levels, or the maximum length of branches, - `LoD[i][j]` is the length of the j-th segment at the i-th level. ## The Offset Representation To quickly access elementary sequences, we adopt an offset representation -- instead of saving the lengths, we save the beginning and ending elements of sequences. In the above example, we accumulate the length of elementary sequences: ``` 3 2 4 1 2 3 ``` into offsets ``` 0 3 5 9 10 12 15 = = = = = = 3 2+3 4+5 1+9 2+10 3+12 ``` so we know that the first sentence is from word 0 to word 3, and the second sentence from work 3 to word 5. Similarly, lengths in the top level LoD ``` 3 1 2 ``` is transformed into offsets of elements/words: ``` 0 9 10 15 = = = 3+2+4 1+9 2+3+10 ``` so we can tell that the first article is from word 0 to word 9, and the second article is from word 9 to word 10. The complete offset representation is as follows: ``` 0 9 10 15 0 3 5 9 10 12 15 ||| || |||| | || ||| ``` ## Slicing of LoD Tensors When we use the above 2-level LoD Tensor as the input to a nested-RNN, we need to retrieve certain sequences. Here we define the sequence identified by branch as the **-slice**. For example, the <2>-slice of above example is ``` 10 15 10 12 15 || ||| ``` and the <2,0>-slice of above slice is ``` 10 12 || ```