/* Copyright (c) 2018 PaddlePaddle Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. */ #ifdef POOL_OP #if defined(__ARM_NEON) || defined(__ARM_NEON__) #include #include "operators/math/pooling.h" namespace paddle_mobile { namespace operators { namespace math { #define POOLING3X3_NORMAL_BORDER(start, end) \ for (int w = start; w < end; ++w) { \ const int w_in_start = -padding_w + w * Stride; \ const int w_in_end = w_in_start + 3; \ const int w_start = w_in_start > 0 ? w_in_start : 0; \ const int w_end = w_in_end < input_w ? w_in_end : input_w; \ PoolingVal

val; \ for (int h_in = h_start; h_in < h_end; ++h_in) { \ for (int w_in = w_start; w_in < w_end; ++w_in) { \ val += input[h_in * input_w + w_in]; \ } \ } \ output_ptr[w] = val.Value(); \ } template struct Pooling3x3NormalRowLoadInput { inline void operator()(const float *input, float32x4x2_t &x0, // NOLINT float32x4x2_t &x1, float32x4x2_t &x2, // NOLINT float32x4x2_t &y0) { // NOLINT x0.val[0] = vld1q_f32(input); x0.val[1] = vld1q_f32(input + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPreq_f32

(x1.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x1.val[1], y0.val[1]); y0.val[0] = vPoolPreq_f32

(x2.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x2.val[1], y0.val[1]); } }; template struct Pooling3x3NormalRowLoadInput { inline void operator()(const float *input, float32x4x2_t &x0, // NOLINT float32x4x2_t &x1, float32x4x2_t &x2, // NOLINT float32x4x2_t &y0) { // NOLINT x0 = vld2q_f32(input); x1 = vld2q_f32(input + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); } }; template inline void Pooling3x3NormalRow(const float *input, const int h_output, const int input_h, const int input_w, const int padding_h, const int padding_w, const int output_w, float *output) { const int h_in_start = -padding_h + h_output * Stride; const int h_in_end = h_in_start + 3; const int h_start = h_in_start > 0 ? h_in_start : 0; const int h_end = h_in_end < input_h ? h_in_end : input_h; float *output_ptr = output + h_output * output_w; if (h_end - h_start <= 0) { memset(output_ptr, 0, output_w * sizeof(float)); return; } const int valid_w_start = (padding_w + Stride - 1) / Stride; const int valid_w_end = (input_w + padding_w - 3) / Stride + 1; const int valid_w = valid_w_end - valid_w_start; // border left POOLING3X3_NORMAL_BORDER(0, valid_w_start) // middle int output_tiles = (valid_w_end - valid_w_start) / 6; int output_tiles_w = output_tiles * 6; Pooling3x3NormalRowLoadInput PoolingCompute; float32x4x2_t x0, x1, x2, y0; float32x4_t post = vdupq_n_f32(1.f / (3 * (h_end - h_start))); for (int w = 0; w < output_tiles_w; w += 6) { int output_offset = valid_w_start + w; int input_w_offset = output_offset * Stride - padding_w; y0.val[0] = vPoolInitq_f32

(); y0.val[1] = vPoolInitq_f32

(); for (int h_in = h_start; h_in < h_end; ++h_in) { PoolingCompute(input + h_in * input_w + input_w_offset, x0, x1, x2, y0); } y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr + output_offset, y0.val[0]); vst1_f32(output_ptr + output_offset + 4, vget_low_f32(y0.val[1])); } int remain = valid_w - output_tiles_w; if (remain > 0) { int remain_start = valid_w_start + output_tiles_w; int input_w_offset = remain_start * Stride - padding_w; float *output_ptr0 = output_ptr + remain_start; y0.val[0] = vPoolInitq_f32

(); y0.val[1] = vPoolInitq_f32

(); for (int h_in = h_start; h_in < h_end; ++h_in) { PoolingCompute(input + h_in * input_w + input_w_offset, x0, x1, x2, y0); } y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); switch (remain) { case 1: vst1q_lane_f32(output_ptr0, y0.val[0], 0); break; case 2: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); break; case 3: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1q_lane_f32(output_ptr0 + 2, y0.val[0], 2); break; case 4: vst1q_f32(output_ptr0, y0.val[0]); break; case 5: vst1q_f32(output_ptr0, y0.val[0]); vst1q_lane_f32(output_ptr0 + 4, y0.val[1], 0); break; } } // border right POOLING3X3_NORMAL_BORDER(valid_w_end, output_w) } template struct Pooling3x3 { inline void operator()(const framework::Tensor &input, const std::vector &paddings, framework::Tensor *output) { const float *input_data = input.data(); float *output_data = output->mutable_data(); int input_h = input.dims()[2]; int input_w = input.dims()[3]; int output_h = output->dims()[2]; int output_w = output->dims()[3]; int padding_h = paddings[0]; int padding_w = paddings[1]; int image_size = input_h * input_w; int out_image_size = output_h * output_w; int valid_h_start = padding_h; int valid_h = input_h - 2; int valid_h_end = valid_h_start + valid_h; int valid_w_start = padding_w; int valid_w = input_w - 2; int valid_w_end = valid_w_start + valid_w; #pragma omp parallel for collapse(2) for (int batch = 0; batch < output->dims()[0]; ++batch) { for (int c = 0; c < output->dims()[1]; ++c) { int channel = batch * output->dims()[1] + c; const float *input_ptr = input_data + channel * image_size; float *output_ptr = output_data + channel * out_image_size; // top for (int h = 0; h < valid_h_start; ++h) { Pooling3x3NormalRow(input_ptr, h, input_h, input_w, padding_h, padding_w, output_w, output_ptr); } // valid int output_w_tiles = valid_w / 6; int output_w_remain = valid_w - output_w_tiles * 6; for (int h = valid_h_start; h < valid_h_end - 3; h += 4) { const float *input_ptr0 = input_ptr + (h - padding_h) * input_w; const float *input_ptr1 = input_ptr0 + input_w; const float *input_ptr2 = input_ptr1 + input_w; const float *input_ptr3 = input_ptr2 + input_w; const float *input_ptr4 = input_ptr3 + input_w; const float *input_ptr5 = input_ptr4 + input_w; float *output_ptr0 = output_ptr + h * output_w; float *output_ptr1 = output_ptr0 + output_w; float *output_ptr2 = output_ptr1 + output_w; float *output_ptr3 = output_ptr2 + output_w; // pad left if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t row3 = vld1_f32(input_ptr3); float32x2_t row4 = vld1_f32(input_ptr4); float32x2_t row5 = vld1_f32(input_ptr5); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, acc1, acc2, acc3, acc12, acc34, post; for (int w = valid_w_start - 1; w >= 0; --w) { int padding = padding_w - w; if (padding >= 3) { output_ptr0[w] = 0.f; output_ptr1[w] = 0.f; output_ptr2[w] = 0.f; output_ptr3[w] = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc12 = vPoolPre_f32

(row1, row2); acc34 = vPoolPre_f32

(row3, row4); acc0 = vPoolPre_f32

(row0, acc12); acc1 = vPoolPre_f32

(row3, acc12); acc2 = vPoolPre_f32

(row2, acc34); acc3 = vPoolPre_f32

(row5, acc34); acc0 = vpPoolPre_f32

(acc0, acc0); acc1 = vpPoolPre_f32

(acc1, acc1); acc2 = vpPoolPre_f32

(acc2, acc2); acc3 = vpPoolPre_f32

(acc3, acc3); acc0 = vPoolPost_f32

(acc0, post); acc1 = vPoolPost_f32

(acc1, post); acc2 = vPoolPost_f32

(acc2, post); acc3 = vPoolPost_f32

(acc3, post); vst1_lane_f32(output_ptr0 + w, acc0, 0); vst1_lane_f32(output_ptr1 + w, acc1, 0); vst1_lane_f32(output_ptr2 + w, acc2, 0); vst1_lane_f32(output_ptr3 + w, acc3, 0); row0 = vext_f32(pad0, row0, 1); row1 = vext_f32(pad0, row1, 1); row2 = vext_f32(pad0, row2, 1); row3 = vext_f32(pad0, row3, 1); row4 = vext_f32(pad0, row4, 1); row5 = vext_f32(pad0, row5, 1); } } output_ptr0 += valid_w_start; output_ptr1 += valid_w_start; output_ptr2 += valid_w_start; output_ptr3 += valid_w_start; } // valid float32x4x2_t x0, x1, x2; float32x4x2_t y0, y1, y2; float32x4_t post = vdupq_n_f32(1.f / 9); for (int loop = 0; loop < output_w_tiles; ++loop) { x0.val[0] = vld1q_f32(input_ptr0); x0.val[1] = vld1q_f32(input_ptr0 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0.val[0] = vld1q_f32(input_ptr1); x0.val[1] = vld1q_f32(input_ptr1 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(y1.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y1.val[1], y0.val[1]); x0.val[0] = vld1q_f32(input_ptr2); x0.val[1] = vld1q_f32(input_ptr2 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y2.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y2.val[1], y1.val[1]); y0.val[0] = vPoolPreq_f32

(y2.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y2.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr0, y0.val[0]); vst1_f32(output_ptr0 + 4, vget_low_f32(y0.val[1])); x0.val[0] = vld1q_f32(input_ptr3); x0.val[1] = vld1q_f32(input_ptr3 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y0.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y0.val[1], y1.val[1]); y2.val[0] = vPoolPreq_f32

(y0.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(y0.val[1], y2.val[1]); y1.val[0] = vPoolPostq_f32

(y1.val[0], post); y1.val[1] = vPoolPostq_f32

(y1.val[1], post); vst1q_f32(output_ptr1, y1.val[0]); vst1_f32(output_ptr1 + 4, vget_low_f32(y1.val[1])); x0.val[0] = vld1q_f32(input_ptr4); x0.val[1] = vld1q_f32(input_ptr4 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], y2.val[1]); y2.val[0] = vPoolPostq_f32

(y2.val[0], post); y2.val[1] = vPoolPostq_f32

(y2.val[1], post); vst1q_f32(output_ptr2, y2.val[0]); vst1_f32(output_ptr2 + 4, vget_low_f32(y2.val[1])); x0.val[0] = vld1q_f32(input_ptr5); x0.val[1] = vld1q_f32(input_ptr5 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr3, y0.val[0]); vst1_f32(output_ptr3 + 4, vget_low_f32(y0.val[1])); input_ptr0 += 6; input_ptr1 += 6; input_ptr2 += 6; input_ptr3 += 6; input_ptr4 += 6; input_ptr5 += 6; output_ptr0 += 6; output_ptr1 += 6; output_ptr2 += 6; output_ptr3 += 6; } // remain width if (output_w_remain > 0) { float32x4x2_t y3; x0.val[0] = vld1q_f32(input_ptr0); x0.val[1] = vld1q_f32(input_ptr0 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0.val[0] = vld1q_f32(input_ptr1); x0.val[1] = vld1q_f32(input_ptr1 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(y1.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y1.val[1], y0.val[1]); x0.val[0] = vld1q_f32(input_ptr2); x0.val[1] = vld1q_f32(input_ptr2 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y2.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y2.val[1], y1.val[1]); y0.val[0] = vPoolPreq_f32

(y2.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y2.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); x0.val[0] = vld1q_f32(input_ptr3); x0.val[1] = vld1q_f32(input_ptr3 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y3.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y3.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y3.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y3.val[1], y1.val[1]); y2.val[0] = vPoolPreq_f32

(y3.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(y3.val[1], y2.val[1]); y1.val[0] = vPoolPostq_f32

(y1.val[0], post); y1.val[1] = vPoolPostq_f32

(y1.val[1], post); x0.val[0] = vld1q_f32(input_ptr4); x0.val[1] = vld1q_f32(input_ptr4 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y3.val[0] = vPoolPreq_f32

(x0.val[0], y3.val[0]); y3.val[1] = vPoolPreq_f32

(x0.val[1], y3.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], y2.val[1]); y2.val[0] = vPoolPostq_f32

(y2.val[0], post); y2.val[1] = vPoolPostq_f32

(y2.val[1], post); x0.val[0] = vld1q_f32(input_ptr5); x0.val[1] = vld1q_f32(input_ptr5 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y3.val[0] = vPoolPreq_f32

(x0.val[0], y3.val[0]); y3.val[1] = vPoolPreq_f32

(x0.val[1], y3.val[1]); y3.val[0] = vPoolPostq_f32

(y3.val[0], post); y3.val[1] = vPoolPostq_f32

(y3.val[1], post); switch (output_w_remain) { case 1: vst1q_lane_f32(output_ptr0, y0.val[0], 0); vst1q_lane_f32(output_ptr1, y1.val[0], 0); vst1q_lane_f32(output_ptr2, y2.val[0], 0); vst1q_lane_f32(output_ptr3, y3.val[0], 0); break; case 2: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1_f32(output_ptr1, vget_low_f32(y1.val[0])); vst1_f32(output_ptr2, vget_low_f32(y2.val[0])); vst1_f32(output_ptr3, vget_low_f32(y3.val[0])); break; case 3: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1_f32(output_ptr1, vget_low_f32(y1.val[0])); vst1_f32(output_ptr2, vget_low_f32(y2.val[0])); vst1_f32(output_ptr3, vget_low_f32(y3.val[0])); vst1q_lane_f32(output_ptr0 + 2, y0.val[0], 2); vst1q_lane_f32(output_ptr1 + 2, y1.val[0], 2); vst1q_lane_f32(output_ptr2 + 2, y2.val[0], 2); vst1q_lane_f32(output_ptr3 + 2, y3.val[0], 2); break; case 4: vst1q_f32(output_ptr0, y0.val[0]); vst1q_f32(output_ptr1, y1.val[0]); vst1q_f32(output_ptr2, y2.val[0]); vst1q_f32(output_ptr3, y3.val[0]); break; case 5: vst1q_f32(output_ptr0, y0.val[0]); vst1q_f32(output_ptr1, y1.val[0]); vst1q_f32(output_ptr2, y2.val[0]); vst1q_f32(output_ptr3, y3.val[0]); vst1q_lane_f32(output_ptr0 + 4, y0.val[1], 0); vst1q_lane_f32(output_ptr1 + 4, y1.val[1], 0); vst1q_lane_f32(output_ptr2 + 4, y2.val[1], 0); vst1q_lane_f32(output_ptr3 + 4, y3.val[1], 0); break; } input_ptr0 += output_w_remain; input_ptr1 += output_w_remain; input_ptr2 += output_w_remain; input_ptr3 += output_w_remain; input_ptr4 += output_w_remain; input_ptr5 += output_w_remain; output_ptr0 += output_w_remain; output_ptr1 += output_w_remain; output_ptr2 += output_w_remain; output_ptr3 += output_w_remain; } // pad right if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t row3 = vld1_f32(input_ptr3); float32x2_t row4 = vld1_f32(input_ptr4); float32x2_t row5 = vld1_f32(input_ptr5); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, acc1, acc2, acc3, acc12, acc34, post; for (int w = valid_w_end; w < output_w; ++w) { int padding = w + 3 - (padding_w + input_w); if (padding >= 3) { *output_ptr0 = 0.f; *output_ptr1 = 0.f; *output_ptr2 = 0.f; *output_ptr3 = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc12 = vPoolPre_f32

(row1, row2); acc34 = vPoolPre_f32

(row3, row4); acc0 = vPoolPre_f32

(row0, acc12); acc1 = vPoolPre_f32

(row3, acc12); acc2 = vPoolPre_f32

(row2, acc34); acc3 = vPoolPre_f32

(row5, acc34); acc0 = vpPoolPre_f32

(acc0, acc0); acc1 = vpPoolPre_f32

(acc1, acc1); acc2 = vpPoolPre_f32

(acc2, acc2); acc3 = vpPoolPre_f32

(acc3, acc3); acc0 = vPoolPost_f32

(acc0, post); acc1 = vPoolPost_f32

(acc1, post); acc2 = vPoolPost_f32

(acc2, post); acc3 = vPoolPost_f32

(acc3, post); vst1_lane_f32(output_ptr0, acc0, 0); vst1_lane_f32(output_ptr1, acc1, 0); vst1_lane_f32(output_ptr2, acc2, 0); vst1_lane_f32(output_ptr3, acc3, 0); row0 = vext_f32(row0, pad0, 1); row1 = vext_f32(row1, pad0, 1); row2 = vext_f32(row2, pad0, 1); row3 = vext_f32(row3, pad0, 1); row4 = vext_f32(row4, pad0, 1); row5 = vext_f32(row5, pad0, 1); } output_ptr0++; output_ptr1++; output_ptr2++; output_ptr3++; } } } // remain height int start_h = valid_h_start + (valid_h & 0xFFFFFFFC); for (int h = start_h; h < valid_h_end; ++h) { const float *input_ptr0 = input_ptr + (h - padding_h) * input_w; const float *input_ptr1 = input_ptr0 + input_w; const float *input_ptr2 = input_ptr1 + input_w; float *output_ptr0 = output_ptr + h * output_w; // pad left if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, post; for (int w = valid_w_start - 1; w >= 0; --w) { int padding = padding_w - w; if (padding >= 3) { output_ptr0[w] = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc0 = vPoolPre_f32

(acc0, row2); acc0 = vpPoolPre_f32

(acc0, acc0); acc0 = vPoolPost_f32

(acc0, post); vst1_lane_f32(output_ptr0 + w, acc0, 0); row0 = vext_f32(pad0, row0, 1); row1 = vext_f32(pad0, row1, 1); row2 = vext_f32(pad0, row2, 1); } } output_ptr0 += valid_w_start; } // valid float32x4x2_t x0, x1, x2, y0; float32x4_t post = vdupq_n_f32(1.f / 9); for (int loop = 0; loop < output_w_tiles; ++loop) { x0.val[0] = vld1q_f32(input_ptr0); x0.val[1] = vld1q_f32(input_ptr0 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0.val[0] = vld1q_f32(input_ptr1); x0.val[1] = vld1q_f32(input_ptr1 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0.val[0] = vld1q_f32(input_ptr2); x0.val[1] = vld1q_f32(input_ptr2 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr0, y0.val[0]); vst1_f32(output_ptr0 + 4, vget_low_f32(y0.val[1])); input_ptr0 += 6; input_ptr1 += 6; input_ptr2 += 6; output_ptr0 += 6; } // remain width if (output_w_remain > 0) { x0.val[0] = vld1q_f32(input_ptr0); x0.val[1] = vld1q_f32(input_ptr0 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0.val[0] = vld1q_f32(input_ptr1); x0.val[1] = vld1q_f32(input_ptr1 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0.val[0] = vld1q_f32(input_ptr2); x0.val[1] = vld1q_f32(input_ptr2 + 4); x1.val[0] = vextq_f32(x0.val[0], x0.val[1], 1); x1.val[1] = vextq_f32(x0.val[1], x0.val[1], 1); x2.val[0] = vextq_f32(x0.val[0], x0.val[1], 2); x2.val[1] = vextq_f32(x0.val[1], x0.val[1], 2); x0.val[0] = vPoolPreq_f32

(x0.val[0], x1.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); // restore switch (output_w_remain) { case 1: vst1q_lane_f32(output_ptr0, y0.val[0], 0); break; case 2: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); break; case 3: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1q_lane_f32(output_ptr0 + 2, y0.val[0], 2); break; case 4: vst1q_f32(output_ptr0, y0.val[0]); break; case 5: vst1q_f32(output_ptr0, y0.val[0]); vst1q_lane_f32(output_ptr0 + 4, y0.val[1], 0); break; } input_ptr0 += output_w_remain; input_ptr1 += output_w_remain; input_ptr2 += output_w_remain; output_ptr0 += output_w_remain; } // pad right if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, post; for (int w = valid_w_end; w < output_w; ++w) { int padding = w + 3 - (padding_w + input_w); if (padding >= 3) { *output_ptr0 = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc0 = vPoolPre_f32

(acc0, row2); acc0 = vpPoolPre_f32

(acc0, acc0); acc0 = vPoolPost_f32

(acc0, post); vst1_lane_f32(output_ptr0, acc0, 0); row0 = vext_f32(row0, pad0, 1); row1 = vext_f32(row1, pad0, 1); row2 = vext_f32(row2, pad0, 1); } output_ptr0++; } } } // pad bottom for (int h = valid_h_end; h < output_h; ++h) { Pooling3x3NormalRow(input_ptr, h, input_h, input_w, padding_h, padding_w, output_w, output_ptr); } } } } }; template struct Pooling3x3 { inline void operator()(const framework::Tensor &input, const std::vector &paddings, framework::Tensor *output) { const float *input_data = input.data(); float *output_data = output->mutable_data(); int input_h = input.dims()[2]; int input_w = input.dims()[3]; int output_h = output->dims()[2]; int output_w = output->dims()[3]; int padding_h = paddings[0]; int padding_w = paddings[1]; int image_size = input_h * input_w; int out_image_size = output_h * output_w; int valid_h_start = (padding_h + 1) / 2; int valid_h_end = (input_h + padding_h - 1) / 2; int valid_h = valid_h_end - valid_h_start; int valid_w_start = (padding_w + 1) / 2; int valid_w_end = (input_w + padding_w - 1) / 2; int valid_w = valid_w_end - valid_w_start; int padding_height = input_h + 2 * padding_h; int padding_width = input_w + 2 * padding_w; bool ceil_mode = (((padding_height - 1) / 2) < output_h) || (((padding_width - 1) / 2) < output_w); int padding_b = padding_h + (ceil_mode ? 2 * output_h - (padding_height - 1) : 0); int padding_r = padding_w + (ceil_mode ? 2 * output_w - (padding_width - 1) : 0); // for pad left int valid_input_w_start = (valid_w_start << 1) - padding_w; #pragma omp parallel for collapse(2) for (int batch = 0; batch < output->dims()[0]; ++batch) { for (int c = 0; c < output->dims()[1]; ++c) { int channel = batch * output->dims()[1] + c; const float *input_ptr = input_data + channel * image_size; float *output_ptr = output_data + channel * out_image_size; // top for (int h = 0; h < valid_h_start; ++h) { Pooling3x3NormalRow(input_ptr, h, input_h, input_w, padding_h, padding_w, output_w, output_ptr); } // valid int output_w_tiles = valid_w / 6; int output_w_remain = valid_w - output_w_tiles * 6; for (int h = valid_h_start; h < valid_h_end - 2; h += 3) { const float *input_ptr0 = input_ptr + (2 * h - padding_h) * input_w; const float *input_ptr1 = input_ptr0 + input_w; const float *input_ptr2 = input_ptr1 + input_w; const float *input_ptr3 = input_ptr2 + input_w; const float *input_ptr4 = input_ptr3 + input_w; const float *input_ptr5 = input_ptr4 + input_w; const float *input_ptr6 = input_ptr5 + input_w; float *output_ptr0 = output_ptr + h * output_w; float *output_ptr1 = output_ptr0 + output_w; float *output_ptr2 = output_ptr1 + output_w; // pad left if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t row3 = vld1_f32(input_ptr3); float32x2_t row4 = vld1_f32(input_ptr4); float32x2_t row5 = vld1_f32(input_ptr5); float32x2_t row6 = vld1_f32(input_ptr6); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, acc1, acc2, post; for (int w = valid_w_start - 1; w >= 0; --w) { int padding = padding_w - (w << 1); if (padding >= 3) { output_ptr0[w] = 0.f; output_ptr1[w] = 0.f; output_ptr2[w] = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc1 = vPoolPre_f32

(row2, row3); acc2 = vPoolPre_f32

(row4, row5); acc0 = vPoolPre_f32

(acc0, row2); acc1 = vPoolPre_f32

(acc1, row4); acc2 = vPoolPre_f32

(acc2, row6); if (padding == 1) { acc0 = vpPoolPre_f32

(acc0, acc0); acc1 = vpPoolPre_f32

(acc1, acc1); acc2 = vpPoolPre_f32

(acc2, acc2); } acc0 = vPoolPost_f32

(acc0, post); acc1 = vPoolPost_f32

(acc1, post); acc2 = vPoolPost_f32

(acc2, post); vst1_lane_f32(output_ptr0 + w, acc0, 0); vst1_lane_f32(output_ptr1 + w, acc1, 0); vst1_lane_f32(output_ptr2 + w, acc2, 0); } } input_ptr0 += valid_input_w_start; input_ptr1 += valid_input_w_start; input_ptr2 += valid_input_w_start; input_ptr3 += valid_input_w_start; input_ptr4 += valid_input_w_start; input_ptr5 += valid_input_w_start; input_ptr6 += valid_input_w_start; output_ptr0 += valid_w_start; output_ptr1 += valid_w_start; output_ptr2 += valid_w_start; } // valid float32x4x2_t x0, x1, x2; float32x4x2_t y0, y1, y2; float32x4_t post = vdupq_n_f32(1.f / 9); for (int loop = 0; loop < output_w_tiles; ++loop) { x0 = vld2q_f32(input_ptr0); x1 = vld2q_f32(input_ptr0 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0 = vld2q_f32(input_ptr1); x1 = vld2q_f32(input_ptr1 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0 = vld2q_f32(input_ptr2); x1 = vld2q_f32(input_ptr2 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(y1.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y1.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr0, y0.val[0]); vst1_f32(output_ptr0 + 4, vget_low_f32(y0.val[1])); x0 = vld2q_f32(input_ptr3); x1 = vld2q_f32(input_ptr3 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], y1.val[1]); x0 = vld2q_f32(input_ptr4); x1 = vld2q_f32(input_ptr4 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y0.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y0.val[1], y1.val[1]); y1.val[0] = vPoolPostq_f32

(y1.val[0], post); y1.val[1] = vPoolPostq_f32

(y1.val[1], post); vst1q_f32(output_ptr1, y1.val[0]); vst1_f32(output_ptr1 + 4, vget_low_f32(y1.val[1])); x0 = vld2q_f32(input_ptr5); x1 = vld2q_f32(input_ptr5 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0 = vld2q_f32(input_ptr6); x1 = vld2q_f32(input_ptr6 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr2, y0.val[0]); vst1_f32(output_ptr2 + 4, vget_low_f32(y0.val[1])); input_ptr0 += 12; input_ptr1 += 12; input_ptr2 += 12; input_ptr3 += 12; input_ptr4 += 12; input_ptr5 += 12; input_ptr6 += 12; output_ptr0 += 6; output_ptr1 += 6; output_ptr2 += 6; } // remain width if (output_w_remain > 0) { x0 = vld2q_f32(input_ptr0); x1 = vld2q_f32(input_ptr0 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0 = vld2q_f32(input_ptr1); x1 = vld2q_f32(input_ptr1 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0 = vld2q_f32(input_ptr2); x1 = vld2q_f32(input_ptr2 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(y1.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(y1.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); x0 = vld2q_f32(input_ptr3); x1 = vld2q_f32(input_ptr3 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(x0.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(x0.val[1], y1.val[1]); x0 = vld2q_f32(input_ptr4); x1 = vld2q_f32(input_ptr4 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y1.val[0] = vPoolPreq_f32

(y2.val[0], y1.val[0]); y1.val[1] = vPoolPreq_f32

(y2.val[1], y1.val[1]); y1.val[0] = vPoolPostq_f32

(y1.val[0], post); y1.val[1] = vPoolPostq_f32

(y1.val[1], post); x0 = vld2q_f32(input_ptr5); x1 = vld2q_f32(input_ptr5 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], y2.val[1]); x0 = vld2q_f32(input_ptr6); x1 = vld2q_f32(input_ptr6 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y2.val[0] = vPoolPreq_f32

(x0.val[0], y2.val[0]); y2.val[1] = vPoolPreq_f32

(x0.val[1], y2.val[1]); y2.val[0] = vPoolPostq_f32

(y2.val[0], post); y2.val[1] = vPoolPostq_f32

(y2.val[1], post); switch (output_w_remain) { case 1: vst1q_lane_f32(output_ptr0, y0.val[0], 0); vst1q_lane_f32(output_ptr1, y1.val[0], 0); vst1q_lane_f32(output_ptr2, y2.val[0], 0); break; case 2: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1_f32(output_ptr1, vget_low_f32(y1.val[0])); vst1_f32(output_ptr2, vget_low_f32(y2.val[0])); break; case 3: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1_f32(output_ptr1, vget_low_f32(y1.val[0])); vst1_f32(output_ptr2, vget_low_f32(y2.val[0])); vst1q_lane_f32(output_ptr0 + 2, y0.val[0], 2); vst1q_lane_f32(output_ptr1 + 2, y1.val[0], 2); vst1q_lane_f32(output_ptr2 + 2, y2.val[0], 2); break; case 4: vst1q_f32(output_ptr0, y0.val[0]); vst1q_f32(output_ptr1, y1.val[0]); vst1q_f32(output_ptr2, y2.val[0]); break; case 5: vst1q_f32(output_ptr0, y0.val[0]); vst1q_f32(output_ptr1, y1.val[0]); vst1q_f32(output_ptr2, y2.val[0]); vst1q_lane_f32(output_ptr0 + 4, y0.val[1], 0); vst1q_lane_f32(output_ptr1 + 4, y1.val[1], 0); vst1q_lane_f32(output_ptr2 + 4, y2.val[1], 0); break; } input_ptr0 += (output_w_remain << 1); input_ptr1 += (output_w_remain << 1); input_ptr2 += (output_w_remain << 1); input_ptr3 += (output_w_remain << 1); input_ptr4 += (output_w_remain << 1); input_ptr5 += (output_w_remain << 1); input_ptr6 += (output_w_remain << 1); output_ptr0 += output_w_remain; output_ptr1 += output_w_remain; output_ptr2 += output_w_remain; } // pad right if (padding_r > 0) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t row3 = vld1_f32(input_ptr3); float32x2_t row4 = vld1_f32(input_ptr4); float32x2_t row5 = vld1_f32(input_ptr5); float32x2_t row6 = vld1_f32(input_ptr6); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, acc1, acc2, post; for (int w = valid_w_end; w < output_w; ++w) { int padding = 2 * w + 3 - (padding_w + input_w); if (padding >= 3) { *output_ptr0 = 0.f; *output_ptr1 = 0.f; *output_ptr2 = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc1 = vPoolPre_f32

(row2, row3); acc2 = vPoolPre_f32

(row4, row5); acc0 = vPoolPre_f32

(acc0, row2); acc1 = vPoolPre_f32

(acc1, row4); acc2 = vPoolPre_f32

(acc2, row6); if (padding == 1) { acc0 = vpPoolPre_f32

(acc0, acc0); acc1 = vpPoolPre_f32

(acc1, acc1); acc2 = vpPoolPre_f32

(acc2, acc2); } acc0 = vPoolPost_f32

(acc0, post); acc1 = vPoolPost_f32

(acc1, post); acc2 = vPoolPost_f32

(acc2, post); vst1_lane_f32(output_ptr0, acc0, 0); vst1_lane_f32(output_ptr1, acc1, 0); vst1_lane_f32(output_ptr2, acc2, 0); } output_ptr0++; output_ptr1++; output_ptr2++; } } } // remain height int start_h = valid_h_start + valid_h / 3 * 3; for (int h = start_h; h < valid_h_end; ++h) { const float *input_ptr0 = input_ptr + (2 * h - padding_h) * input_w; const float *input_ptr1 = input_ptr0 + input_w; const float *input_ptr2 = input_ptr1 + input_w; float *output_ptr0 = output_ptr + h * output_w; // pad left if (padding_w) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, post; for (int w = valid_w_start - 1; w >= 0; --w) { int padding = padding_w - (w << 1); if (padding >= 3) { output_ptr0[w] = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc0 = vPoolPre_f32

(acc0, row2); if (padding == 1) { acc0 = vpPoolPre_f32

(acc0, acc0); } acc0 = vPoolPost_f32

(acc0, post); vst1_lane_f32(output_ptr0 + w, acc0, 0); } } input_ptr0 += valid_input_w_start; input_ptr1 += valid_input_w_start; input_ptr2 += valid_input_w_start; output_ptr0 += valid_w_start; } // valid float32x4x2_t x0, x1, x2, y0; float32x4_t post = vdupq_n_f32(1.f / 9); for (int loop = 0; loop < output_w_tiles; ++loop) { x0 = vld2q_f32(input_ptr0); x1 = vld2q_f32(input_ptr0 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0 = vld2q_f32(input_ptr1); x1 = vld2q_f32(input_ptr1 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0 = vld2q_f32(input_ptr2); x1 = vld2q_f32(input_ptr2 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); vst1q_f32(output_ptr0, y0.val[0]); vst1_f32(output_ptr0 + 4, vget_low_f32(y0.val[1])); input_ptr0 += 12; input_ptr1 += 12; input_ptr2 += 12; output_ptr0 += 6; } // remain width if (output_w_remain > 0) { x0 = vld2q_f32(input_ptr0); x1 = vld2q_f32(input_ptr0 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); x0 = vld2q_f32(input_ptr1); x1 = vld2q_f32(input_ptr1 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); x0 = vld2q_f32(input_ptr2); x1 = vld2q_f32(input_ptr2 + 8); x2.val[0] = vextq_f32(x0.val[0], x1.val[0], 1); x2.val[1] = vextq_f32(x1.val[0], x1.val[0], 1); x0.val[0] = vPoolPreq_f32

(x0.val[0], x0.val[1]); x0.val[1] = vPoolPreq_f32

(x1.val[0], x1.val[1]); x0.val[0] = vPoolPreq_f32

(x0.val[0], x2.val[0]); x0.val[1] = vPoolPreq_f32

(x0.val[1], x2.val[1]); y0.val[0] = vPoolPreq_f32

(x0.val[0], y0.val[0]); y0.val[1] = vPoolPreq_f32

(x0.val[1], y0.val[1]); y0.val[0] = vPoolPostq_f32

(y0.val[0], post); y0.val[1] = vPoolPostq_f32

(y0.val[1], post); // restore switch (output_w_remain) { case 1: vst1q_lane_f32(output_ptr0, y0.val[0], 0); break; case 2: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); break; case 3: vst1_f32(output_ptr0, vget_low_f32(y0.val[0])); vst1q_lane_f32(output_ptr0 + 2, y0.val[0], 2); break; case 4: vst1q_f32(output_ptr0, y0.val[0]); break; case 5: vst1q_f32(output_ptr0, y0.val[0]); vst1q_lane_f32(output_ptr0 + 4, y0.val[1], 0); break; } input_ptr0 += (output_w_remain << 1); input_ptr1 += (output_w_remain << 1); input_ptr2 += (output_w_remain << 1); output_ptr0 += output_w_remain; } // pad right if (padding_r > 0) { float32x2_t row0 = vld1_f32(input_ptr0); float32x2_t row1 = vld1_f32(input_ptr1); float32x2_t row2 = vld1_f32(input_ptr2); float32x2_t pad0 = vPoolInit_f32

(); float32x2_t acc0, post; for (int w = valid_w_end; w < output_w; ++w) { int padding = 2 * w + 3 - (padding_w + input_w); if (padding >= 3) { *output_ptr0 = 0.f; } else { post = vdup_n_f32(1.f / (3 * (3 - padding))); acc0 = vPoolPre_f32

(row0, row1); acc0 = vPoolPre_f32

(acc0, row2); if (padding == 1) { acc0 = vpPoolPre_f32

(acc0, acc0); } acc0 = vPoolPost_f32

(acc0, post); vst1_lane_f32(output_ptr0, acc0, 0); } output_ptr0++; } } } // bottom for (int h = valid_h_end; h < output_h; ++h) { Pooling3x3NormalRow(input_ptr, h, input_h, input_w, padding_h, padding_w, output_w, output_ptr); } } } } }; template struct Pooling3x3; template struct Pooling3x3; template struct Pooling3x3; template struct Pooling3x3; } // namespace math } // namespace operators } // namespace paddle_mobile #endif // __ARM_NEON #endif // POOL_OP