提交 c2a14556 编写于 作者: H hanbuhe

merge fc commit

上级 2e29d186
......@@ -47,7 +47,7 @@ static Dtype find_max(Dtype* data, int64_t num) {
}
// template <typename Dtype>
void quantify_filter(framework::Tensor* filter) {
void quantize_filter(framework::Tensor *filter) {
DLOG << "quantilize_filter........";
float scale = 0;
......
......@@ -24,7 +24,7 @@ template <typename Dtype>
static void chw_to_hwc(Dtype* data_in, Dtype* data_out, int64_t num,
int64_t channel, int64_t height, int64_t width);
void quantify_filter(framework::Tensor* filter);
void quantize_filter(framework::Tensor *filter);
} // namespace fpga
} // namespace paddle_mobile
......@@ -60,7 +60,7 @@ bool ConvAddBNKernel<FPGA, float>::Init(FusionConvAddBNParam *param) {
param->SetNewScale(new_scale);
param->SetNewBias(new_bias);
fpga::quantify_filter(filter);
fpga::quantize_filter(filter);
auto filter_ptr = filter->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -56,7 +56,7 @@ bool ConvAddBNReluKernel<FPGA, float>::Init(FusionConvAddBNReluParam *param) {
}
param->SetNewScale(new_scale);
param->SetNewBias(new_bias);
fpga::quantify_filter(filter);
fpga::quantize_filter(filter);
auto filter_ptr = filter->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -40,7 +40,7 @@ bool ConvAddReluKernel<FPGA, float>::Init(FusionConvAddReluParam *param) {
bs_ptr[i * 2 + 1] = bias_ptr[i];
}
fpga::quantify_filter(filter);
fpga::quantize_filter(filter);
auto filter_ptr = filter->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -55,7 +55,7 @@ bool ConvBNKernel<FPGA, float>::Init(FusionConvBNParam *param) {
}
param->SetNewScale(new_scale);
param->SetNewBias(new_bias);
fpga::quantify_filter(filter);
fpga::quantize_filter(filter);
auto filter_ptr = filter->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -52,7 +52,7 @@ bool ConvBNReluKernel<FPGA, float>::Init(FusionConvBNReluParam *param) {
}
param->SetNewScale(new_scale);
param->SetNewBias(new_bias);
fpga::quantify_filter(filter);
fpga::quantize_filter(filter);
auto filter_ptr = filter->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -40,7 +40,7 @@ bool FusionFcReluKernel<FPGA, float>::Init(FusionFcReluParam *param) {
bs_ptr[i * 2 + 1] = input_z_ptr[i];
}
fpga::quantify_filter(input_y);
fpga::quantize_filter(input_y);
auto input_y_ptr = input_y->data<int8_t>();
fpga::ConvArgs convArgs;
......
......@@ -14,7 +14,7 @@ limitations under the License. */
#ifdef FUSION_FC_OP
#include "operators/kernel/fusion_fc_kernel.h"
#include "fpga/fpga_quantilization.h"
#include "fpga/quantization.h"
namespace paddle_mobile {
namespace operators {
......@@ -39,7 +39,7 @@ bool FusionFcKernel<FPGA, float>::Init(FusionFcParam *param) {
bs_ptr[i * 2 + 1] = input_z_ptr[i];
}
fpga::quantify_filter(input_y);
fpga::quantize_filter(input_y);
auto input_y_ptr = input_y->data<int8_t>();
fpga::ConvArgs convArgs;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册