// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include "lite/core/kernel.h" #include "lite/core/op_lite.h" #include "lite/core/scope.h" #include "lite/core/tensor.h" #include "lite/operators/op_params.h" #include "lite/utils/all.h" namespace paddle { namespace lite { namespace operators { class FakeQuantizeRangeMaxAbsOpLite : public OpLite { public: FakeQuantizeRangeMaxAbsOpLite() {} explicit FakeQuantizeRangeMaxAbsOpLite(const std::string &type) : OpLite(type) {} bool CheckShape() const override { return true; } bool InferShapeImpl() const override { return true; } bool AttachImpl(const cpp::OpDesc &op_desc, lite::Scope *scope) override { auto x = op_desc.Input("X").front(); if (op_desc.HasInput("InScale")) { auto in_scale = op_desc.Input("InScale").front(); param_.in_scale = scope->FindVar(in_scale)->GetMutable(); } auto out = op_desc.Output("Out").front(); auto out_scale = op_desc.Output("OutScale").front(); param_.x = scope->FindVar(x)->GetMutable(); param_.out = scope->FindVar(out)->GetMutable(); param_.out_scale = scope->FindVar(out_scale)->GetMutable(); param_.bit_length = op_desc.GetAttr("bit_length"); return true; } void AttachKernel(KernelBase *kernel) override { kernel->SetParam(param_); } std::string DebugString() const override { return "fake_quantize_range_max_abs"; } private: mutable FakeQuantizeMovingAvgMaxAbsParam param_; }; } // namespace operators } // namespace lite } // namespace paddle