提交 bfc02229 编写于 作者: Z zhangjun

fix

上级 a5113877
......@@ -33,10 +33,10 @@ enum class Precision {
kFloat32 = 0, // fp32
kInt8, // int8
kHalf, // fp16
kBfloat16 // bf16
kBfloat16, // bf16
};
std::string PrecisionTypeString(const Precision data_type) {
static std::string PrecisionTypeString(const Precision data_type) {
switch (data_type) {
case Precision::kFloat32:
return "kFloat32";
......@@ -51,7 +51,7 @@ std::string PrecisionTypeString(const Precision data_type) {
}
}
std::string ToLower(const std::string& data) {
static std::string ToLower(const std::string& data) {
std::string result = data;
std::transform(
result.begin(), result.end(), result.begin(), [](unsigned char c) {
......
......@@ -21,13 +21,12 @@ namespace inference {
DEFINE_int32(gpuid, 0, "GPU device id to use");
DEFINE_string(precision, "fp32", "precision to deploy, default is fp32");
DEFINE_bool(use_calib, false, "calibration mode, default is false")
DEFINE_bool(use_calib, false, "calibration mode, default is false");
REGIST_FACTORY_OBJECT_IMPL_WITH_NAME(
::baidu::paddle_serving::predictor::FluidInferEngine<
PaddleInferenceEngine>,
::baidu::paddle_serving::predictor::InferEngine,
"PADDLE_INFER");
REGIST_FACTORY_OBJECT_IMPL_WITH_NAME(
::baidu::paddle_serving::predictor::FluidInferEngine<PaddleInferenceEngine>,
::baidu::paddle_serving::predictor::InferEngine,
"PADDLE_INFER");
} // namespace inference
} // namespace paddle_serving
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册