Design Doc: PaddlePaddle Fluid

Why Fluid

When Baidu developed PaddlePaddle in 2013, the only well-known open source deep learning system at the time was Caffe. However, when PaddlePaddle was open-sourced in 2016, many other choices were available. There was a challenge – what is the need for open sourcing yet another deep learning framework?

Fluid is the answer. Fluid is similar to PyTorch and TensorFlow Eager Execution, which describes the “process” of training or inference using the concept of a model. In fact in PyTorch, TensorFlow Eager Execution and Fluid, there is no concept of a model at all. The details are covered in the sections below. Fluid is currently more extreme in the above mentioned idea than PyTorch and Eager Execution, and we are trying to push Fluid towards the directions of a compiler and a new programming language for deep learning.

The Evolution of Deep Learning Systems

Deep learning infrastructure is one of the fastest evolving technologies. Within four years, there have already been three generations of technologies invented.

| Existed since | model as sequence of layers | model as graph of operators | No model | |–|–|–|–| | 2013 | Caffe, Theano, Torch, PaddlePaddle | | | | 2015 | | TensorFlow, MxNet, Caffe2, ONNX, n-graph | | | 2016 | | | PyTorch, TensorFlow Eager Execution, PaddlePaddle Fluid |

From the above table, we see that the deep learning technology is evolving towards getting rid of the concept of a model. To understand the reasons behind this direction, a comparison of the programming paradigms or the ways to program deep learning applications using these systems, would be helpful. The following section goes over these.

Deep Learning Programming Paradigms

With the systems listed as the first or second generation, e.g., Caffe or TensorFlow, an AI application training program looks like the following:

x = layer.data("image")
l = layer.data("label")
f = layer.fc(x, W)
s = layer.softmax(f)
c = layer.mse(l, s)

for i in xrange(1000): # train for 1000 iterations
    m = read_minibatch()
    forward({input=x, data=m}, minimize=c)
    backward(...)

print W # print the trained model parameters.

The above program includes two parts:

  1. The first part describes the model, and
  2. The second part describes the training process (or inference process) for the model.

This paradigm has a well-known problem that limits the productivity of programmers. If the programmer made a mistake in configuring the model, the error messages wouldn’t show up until the second part is executed and forward and backward propagations are performed. This makes it difficult for the programmer to debug and locate a mistake that is located blocks away from the actual error prompt.

This problem of being hard to debug and re-iterate fast on a program is the primary reason that programmers, in general, prefer PyTorch over the older systems. Using PyTorch, we would write the above program as following:

W = tensor(...)

for i in xrange(1000): # train for 1000 iterations
    m = read_minibatch()
    x = m["image"]
    l = m["label"]
    f = layer.fc(x, W)
    s = layer.softmax(f)
    c = layer.mse(l, s)
    backward()

print W # print the trained model parameters.

We can see that the main difference is the moving the model configuration part (the first step) into the training loop. This change would allow the mistakes in model configuration to be reported where they actually appear in the programming block. This change also represents the model better, or its forward pass, by keeping the configuration process in the training loop.

Describe Arbitrary Models for the Future

Describing the process instead of the model also brings Fluid, the flexibility to define different non-standard models that haven’t been invented yet.

As we write out the program for the process, we can write an RNN as a loop, instead of an RNN as a layer or as an operator. A PyTorch example would look like the following:

for i in xrange(1000):
    m = read_minibatch()
    x = m["sentence"]
    for t in xrange x.len():
        h[t] = the_step(x[t])

With Fluid, the training loop and the RNN in the above program are not really Python loops, but just a “loop structure” provided by Fluid and implemented in C++ as the following:

train_loop = layers.While(cond)
with train_loop.block():
  m = read_minibatch()
  x = m["sentence"]
  rnn = layers.While(...)
  with rnn.block():
    h[t] = the_step(input[t])

An actual Fluid example is described here.

From the example, the Fluid programs look very similar to their PyTorch equivalent programs, except that Fluid’s loop structure, wrapped with Python’s with statement, could run much faster than just a Python loop.

We have more examples of the if-then-else structure of Fluid.

Turing Completeness

In computability theory, a system of data-manipulation rules, such as a programming language, is said to be Turing complete if it can be used to simulate any Turing machine. For a programming language, if it provides if-then-else and loop, it is Turing complete. From the above examples, Fluid seems to be Turing complete; however, it is noteworthy to notice that there is a slight difference between the if-then-else of Fluid and that of a programming language. The difference being that the former runs both of its branches and splits the input mini-batch into two – one for the True condition and another for the False condition. This hasn’t been researched in depth if this is equivalent to the if-then-else in programming languages that makes them Turing-complete. Based on a conversation with Yuang Yu, it seems to be the case but this needs to be looked into in-depth.

The Execution of a Fluid Program

There are two ways to execute a Fluid program. When a program is executed, it creates a protobuf message ProgramDesc that describes the process and is conceptually like an abstract syntax tree.

There is a C++ class Executor, which runs a ProgramDesc, similar to how an interpreter runs a Python program.

Fluid is moving towards the direction of a compiler, which is explain in fluid.

Backward Compatibility of Fluid

Given all the advantages from the removal of the concept of a model, hardware manufacturers might still prefer the existence of the concept of a model, so it would be easier for them to support multiple frameworks all at once and could run a trained model during inference. For example, Nervana, a startup company acquired by Intel, has been working on an XPU that reads the models in the format known as n-graph. Similarly, Movidius is producing a mobile deep learning chip that reads and runs graphs of operators. The well-known ONNX is also a file format of graphs of operators.

For Fluid, we can write a converter that extracts the parts in the ProgramDesc protobuf message, converts them into a graph of operators, and exports the graph into the ONNX or n-graph format.