// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include #include #include "lite/backends/opencl/cl_image_converter.h" #include "lite/backends/opencl/target_wrapper.h" #include "lite/core/op_registry.h" #include "lite/core/tensor.h" namespace paddle { namespace lite { template void depth_conv(const T* input_data, const lite::DDim& input_dims, const T* filter_data, const lite::DDim& filter_dims, T* output_data, const lite::DDim& output_dims) { int stride_h = STRIDE_H, stride_w = STRIDE_W; int64_t batches = input_dims[0]; int64_t channels = input_dims[1]; int64_t h = input_dims[2]; int64_t w = input_dims[3]; int64_t num_output = output_dims[1]; int64_t outh = output_dims[2]; int64_t outw = output_dims[3]; int64_t filter_h = filter_dims[2]; int64_t filter_w = filter_dims[3]; const int64_t in_batch_size = channels * h * w; const int64_t out_batch_size = num_output * outh * outw; auto kernel_offset = std::unique_ptr(new int[filter_h * filter_w]); { int p = 0; int offset = 0; int gap = w - filter_w; for (int i = 0; i < filter_h; i++) { for (int j = 0; j < filter_w; j++) { kernel_offset[p++] = offset; offset += 1; } offset += gap; } } for (int b = 0; b < batches; b++) { auto* input_batch_start = input_data + b * in_batch_size; auto* output_batch_start = output_data + b * out_batch_size; for (int p = 0; p < num_output; p++) { float* output_ptr = output_batch_start + p * outh * outw; const float* filter_ptr = filter_data + p * filter_h * filter_w; const float* input_ptr = input_batch_start + p * h * w; for (int i = 0; i < outh; i++) { for (int j = 0; j < outw; j++) { float sum = 0; const float* input_ch_start = input_ptr + i * stride_h * w + j * stride_w; for (int fh = 0; fh < filter_h; ++fh) { for (int fw = 0; fw < filter_w; ++fw) { float val = input_ch_start[kernel_offset[fh * filter_w + fw]]; float w = filter_ptr[fh * filter_w + fw]; sum += val * w; } } output_ptr[j] = sum; } output_ptr += outw; } } } } TEST(depthwise_conv2d_buffer_fp32, compute) { LOG(INFO) << "to get kernel ..."; auto kernels = KernelRegistry::Global().Create("depthwise_conv2d", TARGET(kOpenCL), PRECISION(kFloat), DATALAYOUT(kNCHW)); ASSERT_FALSE(kernels.empty()); auto kernel = std::move(kernels.front()); LOG(INFO) << "get kernel"; lite::Tensor input, filter, output; operators::ConvParam param; param.x = &input; param.filter = &filter; param.output = &output; std::vector paddings = {0, 0}; param.paddings = std::make_shared>(paddings); param.strides = std::vector{1, 1}; std::unique_ptr context(new KernelContext); context->As().InitOnce(); kernel->SetParam(param); std::unique_ptr dep_context(new KernelContext); context->As().CopySharedTo( &(dep_context->As())); kernel->SetContext(std::move(dep_context)); std::default_random_engine engine; std::uniform_real_distribution gen(-5, 5); std::vector input_v(4 * 32 * 112 * 112); std::vector filter_v(32 * 1 * 3 * 3); for (auto& i : input_v) { i = gen(engine); } for (auto& f : filter_v) { f = gen(engine); } input.Assign( input_v.data(), lite::DDim{std::vector({4, 32, 112, 112})}); filter.Assign( filter_v.data(), lite::DDim{std::vector({32, 1, 3, 3})}); output.Resize({4, 32, 110, 110}); kernel->Launch(); auto* wait_list = context->As().cl_wait_list(); auto* out_ptr = param.output->data(); auto it = wait_list->find(out_ptr); if (it != wait_list->end()) { VLOG(4) << "--- Find the sync event for the target cl tensor. ---"; auto& event = *(it->second); event.wait(); } else { LOG(FATAL) << "Could not find the sync event for the target cl tensor."; } lite::Tensor output_ref; output_ref.Resize({4, 32, 110, 110}); auto* output_ref_data = output_ref.mutable_data(TARGET(kARM)); auto* input_data = input.mutable_data(); auto* filter_data = filter.mutable_data(); auto* mapped_input = static_cast(TargetWrapperCL::Map( input_data, 0, sizeof(float) * input.dims().production())); auto* mapped_filter = static_cast(TargetWrapperCL::Map( filter_data, 0, sizeof(float) * filter.dims().production())); depth_conv(mapped_input, input.dims(), mapped_filter, filter.dims(), output_ref_data, output_ref.dims()); auto* output_data = output.mutable_data(); auto* mapped_output = static_cast(TargetWrapperCL::Map( output_data, 0, sizeof(float) * output.dims().production())); for (int i = 0; i < output.dims().production(); i++) { EXPECT_NEAR(mapped_output[i], output_ref_data[i], 1e-4); } TargetWrapperCL::Unmap(output_data, mapped_output); TargetWrapperCL::Unmap(filter_data, mapped_filter); TargetWrapperCL::Unmap(input_data, mapped_input); } } // namespace lite } // namespace paddle USE_LITE_KERNEL(depthwise_conv2d, kOpenCL, kFloat, kNCHW, def);