diff --git a/doc/fluid/design/dist_train/prefetch_parameter.md b/doc/fluid/design/dist_train/prefetch_parameter.md
index 952d2bada9238b0668225893d661c4856464b35a..b3e2210fdff26b06cb4fc5e6f6ffdef3ab8b20f8 100644
--- a/doc/fluid/design/dist_train/prefetch_parameter.md
+++ b/doc/fluid/design/dist_train/prefetch_parameter.md
@@ -16,13 +16,13 @@ Prior to reading this design, it would be useful for the reader to make themselv
The execution of `lookup local table` is as follows:
-
+
For some cases, the parameter(`weight`) may be very large, such as 10 billion features, the entire
data could not be stored in one trainer's memory, so we need to partition this parameter and
pre-fetch it at the beginning of each mini-batch, and we call it `lookup remote table`:
-
+
The processing flow of `lookup remote table` is as follows: