diff --git a/paddle/fluid/operators/conv_mkldnn_op.cc b/paddle/fluid/operators/conv_mkldnn_op.cc index 3d437f219991aa3b2161d5521b5714444c9adf51..43aba94dd2b45c763dfa6681b13a3852dc4be325 100644 --- a/paddle/fluid/operators/conv_mkldnn_op.cc +++ b/paddle/fluid/operators/conv_mkldnn_op.cc @@ -278,7 +278,7 @@ class ConvMKLDNNOpKernel : public paddle::framework::OpKernel { auto* bias = ctx.HasInput("Bias") ? ctx.Input("Bias") : nullptr; auto* output = ctx.Output("Output"); - bool is_INT8 = ctx.HasInput("Bias")? true : false; + bool is_INT8 = ctx.HasInput("Scale_in")? true : false; auto* scale_in = ctx.HasInput("Scale_in") ? ctx.Input("Scale_in") : nullptr; auto* scale_in_eltwise = ctx.HasInput("Scale_in_eltwise")? ctx.Input("Scale_in_eltwise") : nullptr; auto* scale_weights = ctx.HasInput("Scale_weights")? ctx.Input("Scale_weights") : nullptr; @@ -478,7 +478,7 @@ class ConvMKLDNNOpKernel : public paddle::framework::OpKernel { if (fuse_relu) { constexpr float scale = 1.0f; constexpr float negative_slope = 0.0f; - constexpr float placeholder = 0.0f; + constexpr float placeholder = 0.0f; //beta post_operations.append_eltwise(scale, mkldnn::algorithm::eltwise_relu, negative_slope, placeholder); } diff --git a/paddle/fluid/operators/dequantization_op.cc b/paddle/fluid/operators/dequantize_op.cc similarity index 68% rename from paddle/fluid/operators/dequantization_op.cc rename to paddle/fluid/operators/dequantize_op.cc index 68afe0825539861ff70402d7efd6b38149cf01a8..1ee5ed61ec0ed6be2d6b32b9cb8d031175ac6ae7 100644 --- a/paddle/fluid/operators/dequantization_op.cc +++ b/paddle/fluid/operators/dequantize_op.cc @@ -16,7 +16,7 @@ limitations under the License. */ #include "mkldnn.hpp" #include "paddle/fluid/framework/tensor.h" #include "paddle/fluid/platform/mkldnn_helper.h" -#include "paddle/fluid/operators/dequantization_op.h" +#include "paddle/fluid/operators/dequantize_op.h" #include "paddle/fluid/framework/data_layout_transform.h" namespace paddle { @@ -36,29 +36,6 @@ template class DeQuantOpKernel : public framework::OpKernel { public: - // MKLDNNDataType ToMKLDNNDataType(const std::type_index type) { - // static const std::map dict{ - // {std::type_index(typeid(float)), MKLDNNDataType::f32}, // NOLINT - // {std::type_index(typeid(char)), MKLDNNDataType::s8}, // NOLINT - // {std::type_index(typeid(unsigned char)), MKLDNNDataType::u8}, - // {std::type_index(typeid(int16_t)), MKLDNNDataType::s16}, - // {std::type_index(typeid(int32_t)), MKLDNNDataType::s32}}; - // auto iter = dict.find(type); - // if (iter != dict.end()) return iter->second; - // return MKLDNNDataType::data_undef; - // } - //mkldnn::memory::data_type ToMKLDNNDataType(const std::type_index type) const override{ - // static const std::map dict{ - // {std::type_index(typeid(float)), mkldnn::memory::data_type::f32}, // NOLINT - // {std::type_index(typeid(char)), mkldnn::memory::data_type::s8}, // NOLINT - // {std::type_index(typeid(unsigned char)), mkldnn::memory::data_type::u8}, - // {std::type_index(typeid(int16_t)), mkldnn::memory::data_type::s16}, - // {std::type_index(typeid(int32_t)), mkldnn::memory::data_type::s32}}; - // auto iter = dict.find(type); - // if (iter != dict.end()) return iter->second; - // return mkldnn::memory::data_type::data_undef; - //} - void Compute(const framework::ExecutionContext& ctx) const override { auto* input = ctx.Input("Input"); auto* scale = ctx.Input("Scale"); @@ -77,15 +54,14 @@ class DeQuantOpKernel : public framework::OpKernel { std::vector src_tz = paddle::framework::vectorize2int(input->dims()); std::vector dst_tz = paddle::framework::vectorize2int(output->dims()); mkldnn::memory::data_type src_dt = paddle::framework::ToMKLDNNDataType(input->type()); - mkldnn::memory::format src_fmt = input->format(); + mkldnn::memory::format src_fmt = memory::format::nhwc;//input->format(); mkldnn::primitive_attr attri; int mask = 0; attri.set_output_scales(mask, scale_data); - //attri.set_int_output_round_mode(round_nearest); //FIX ME auto src_md = platform::MKLDNNMemDesc( - {src_tz}, src_dt, src_fmt); //FIX ME WITH S8 + {src_tz}, src_dt, src_fmt); auto src_pd = mkldnn::memory::primitive_desc{src_md, engine}; auto src_memory = std::make_shared(src_pd, to_void_cast(input_data)); std::shared_ptr src_memory_p = std::shared_ptr(new primitive::at(*src_memory)); @@ -130,8 +106,8 @@ This op will quantize data from INT8 to FP32 namespace ops = paddle::operators; -REGISTER_OPERATOR(dequantization, ops::DeQuantOp, ops::DeQuantOpMaker, paddle::framework::DefaultGradOpDescMaker); +REGISTER_OPERATOR(dequantize, ops::DeQuantOp, ops::DeQuantOpMaker, paddle::framework::DefaultGradOpDescMaker); -REGISTER_OP_CPU_KERNEL(dequantization, ops::DeQuantOpKernel); +REGISTER_OP_CPU_KERNEL(dequantize, ops::DeQuantOpKernel); diff --git a/paddle/fluid/operators/dequantization_op.h b/paddle/fluid/operators/dequantize_op.h similarity index 100% rename from paddle/fluid/operators/dequantization_op.h rename to paddle/fluid/operators/dequantize_op.h diff --git a/paddle/fluid/operators/quantization_op.cc b/paddle/fluid/operators/quantize_op.cc similarity index 92% rename from paddle/fluid/operators/quantization_op.cc rename to paddle/fluid/operators/quantize_op.cc index 17f07beaa6feaf6037636789bae913e150f26f24..06336b025e4cb4ad7886a94597d23aa472041358 100644 --- a/paddle/fluid/operators/quantization_op.cc +++ b/paddle/fluid/operators/quantize_op.cc @@ -16,7 +16,7 @@ limitations under the License. */ #include "mkldnn.hpp" #include "paddle/fluid/framework/tensor.h" #include "paddle/fluid/platform/mkldnn_helper.h" -#include "paddle/fluid/operators/quantization_op.h" +#include "paddle/fluid/operators/quantize_op.h" namespace paddle { namespace operators { @@ -53,7 +53,6 @@ class QuantOpKernel : public framework::OpKernel { mkldnn::primitive_attr attri; int mask = 0; attri.set_output_scales(mask, scale_data); - //attri.set_int_output_round_mode(round_nearest); //FIX ME auto src_md = platform::MKLDNNMemDesc( {src_tz}, memory::data_type::f32, input->format()); @@ -102,9 +101,9 @@ This op will quantize data from FP32 to INT8 namespace ops = paddle::operators; -REGISTER_OPERATOR(quantization, ops::QuantOp, ops::QuantOpMaker, paddle::framework::DefaultGradOpDescMaker); +REGISTER_OPERATOR(quantize, ops::QuantOp, ops::QuantOpMaker, paddle::framework::DefaultGradOpDescMaker); -REGISTER_OP_CPU_KERNEL(quantization, ops::QuantOpKernel); +REGISTER_OP_CPU_KERNEL(quantize, ops::QuantOpKernel); //REGISTER_OP_KERNEL(quantization, MKLDNN, paddle::platform::CPUPlace, ops::QuantOpKernel); diff --git a/paddle/fluid/operators/quantization_op.h b/paddle/fluid/operators/quantize_op.h similarity index 100% rename from paddle/fluid/operators/quantization_op.h rename to paddle/fluid/operators/quantize_op.h diff --git a/paddle/fluid/operators/requantization_op.cc b/paddle/fluid/operators/requantize_op.cc similarity index 90% rename from paddle/fluid/operators/requantization_op.cc rename to paddle/fluid/operators/requantize_op.cc index 9454a7e8dd1169512ad5ffbac31065a024e27e2c..9abc0a4ea45ef8ebcd8a96571823d9dace05fa7f 100644 --- a/paddle/fluid/operators/requantization_op.cc +++ b/paddle/fluid/operators/requantize_op.cc @@ -16,7 +16,7 @@ limitations under the License. */ #include "mkldnn.hpp" #include "paddle/fluid/framework/tensor.h" #include "paddle/fluid/platform/mkldnn_helper.h" -#include "paddle/fluid/operators/requantization_op.h" +#include "paddle/fluid/operators/requantize_op.h" #include "paddle/fluid/framework/data_layout_transform.h" namespace paddle { @@ -48,8 +48,8 @@ class ReQuantOpKernel : public framework::OpKernel { std::vector dst_tz = paddle::framework::vectorize2int(output->dims()); mkldnn::memory::data_type src_dt = paddle::framework::ToMKLDNNDataType(input->type()); mkldnn::memory::data_type dst_dt = paddle::framework::ToMKLDNNDataType(output->type()); - mkldnn::memory::format src_fmt = input->format(); - mkldnn::memory::format dst_fmt = output->format(); + mkldnn::memory::format src_fmt = memory::format::nhwc;//input->format(); + mkldnn::memory::format dst_fmt = memory::format::nhwc;//output->format(); const T* input_data = input->data(); T* output_data = output->mutable_data(ctx.GetPlace()); @@ -107,7 +107,7 @@ This op will requantize data from INT8 to INT8 namespace ops = paddle::operators; -REGISTER_OPERATOR(requantization, ops::ReQuantOp, ops::ReQuantOpMaker, paddle::framework::DefaultGradOpDescMaker); +REGISTER_OPERATOR(requantize, ops::ReQuantOp, ops::ReQuantOpMaker, paddle::framework::DefaultGradOpDescMaker); -REGISTER_OP_CPU_KERNEL(requantization, ops::ReQuantOpKernel); +REGISTER_OP_CPU_KERNEL(requantize, ops::ReQuantOpKernel); diff --git a/paddle/fluid/operators/requantization_op.h b/paddle/fluid/operators/requantize_op.h similarity index 100% rename from paddle/fluid/operators/requantization_op.h rename to paddle/fluid/operators/requantize_op.h