From 219a55cd1c13e6fce43d1f23ba3b80a07d86438f Mon Sep 17 00:00:00 2001 From: Kexin Zhao Date: Mon, 7 May 2018 15:46:22 -0700 Subject: [PATCH] Change the float16 inference report to README.md (#10465) * change to readme * Update README.md --- contrib/float16/{float16_inference_report.md => README.md} | 4 ++++ 1 file changed, 4 insertions(+) rename contrib/float16/{float16_inference_report.md => README.md} (99%) diff --git a/contrib/float16/float16_inference_report.md b/contrib/float16/README.md similarity index 99% rename from contrib/float16/float16_inference_report.md rename to contrib/float16/README.md index 67623a4d8d5..171c222ba8a 100644 --- a/contrib/float16/float16_inference_report.md +++ b/contrib/float16/README.md @@ -1,3 +1,7 @@ +# Float16 Inference in PaddlePaddle Fluid + +Kexin Zhao + ## Introduction Working with deep neural networks (DNN) is a two-stage process. First we train DNN using labeled examples of inputs and desired outputs to obtain the model parameters (weights), then we deploy DNN along with the trained weights to run inference on unknown inputs. Typically, these weights are in float data type and hence we run inference in float mode using these weights. This post focuses on the discussion of how to use low precision float16 data type to represent these trained weights and run inference in float16 mode as well as the advantages of float16 inference over its float counterpart by showing some experiment results. -- GitLab