// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include #include #include "lite/backends/opencl/cl_half.h" #include "lite/backends/opencl/cl_image_converter.h" #include "lite/backends/opencl/cl_include.h" #include "lite/core/kernel.h" #include "lite/core/op_registry.h" #include "lite/kernels/opencl/image_helper.h" #include "lite/operators/op_params.h" #include "lite/utils/logging.h" #include "lite/utils/replace_stl/stream.h" namespace paddle { namespace lite { namespace kernels { namespace opencl { class InstanceNormImageCompute : public KernelLite { public: using param_t = operators::InstanceNormParam; std::string doc() const override { return "InstanceNorm using cl::Image2D(ImageDefault/RGBA), kFP16"; } void PrepareForRun() override { instance_norm_param_ = param_.get_mutable(); auto channel = instance_norm_param_->scale->dims()[0]; auto batch = instance_norm_param_->x->dims()[0]; int64_t cgroup = (channel + 3) / 4; int64_t cround = cgroup * 4; std::vector scale_img(cround * batch); std::vector bias_img(cround * batch); const float* scale_data = instance_norm_param_->scale->data(); const float* bias_data = instance_norm_param_->bias->data(); //! init scale_img bias_img data for (int i = 0; i < channel; ++i) { scale_img[i] = Float2Half(scale_data[i]); bias_img[i] = Float2Half(bias_data[i]); } for (int i = channel; i < cround; ++i) { scale_img[i] = Float2Half(0.f); bias_img[i] = Float2Half(0.f); } for (int i = 1; i < batch; ++i) { memcpy(scale_img.data() + i * cround, scale_img.data(), cround * sizeof(half_t)); memcpy(bias_img.data() + i * cround, bias_img.data(), cround * sizeof(half_t)); } DDim scale_img_size{{cgroup, batch}}; scale_image_.mutable_data( scale_img_size[0], scale_img_size[1], scale_img.data()); bias_image_.mutable_data( scale_img_size[0], scale_img_size[1], bias_img.data()); auto& context = ctx_->As(); context.cl_context()->AddKernel( kernel_func_name_, "image/instance_norm_kernel.cl", build_options_); VLOG(1) << "kernel_func_name_:" << kernel_func_name_; } void Run() override { auto& context = ctx_->As(); CHECK(context.cl_context() != nullptr); auto* x = instance_norm_param_->x; auto* out = instance_norm_param_->out; auto in_dims = x->dims(); int batch = in_dims[0]; int channel = in_dims[1]; int in_h = in_dims[2]; int in_w = in_dims[3]; VLOG(4) << "x->target():" << TargetToStr(x->target()); VLOG(4) << "out->target():" << TargetToStr(out->target()); VLOG(4) << "x->dims():" << in_dims; auto out_image_shape = InitImageDimInfoWith(in_dims); auto* x_img = x->data(); auto* out_img = out->mutable_data( out_image_shape["width"], out_image_shape["height"]); VLOG(4) << "out_image_shape[w,h]: " << out_image_shape["width"] << " " << out_image_shape["height"]; VLOG(4) << "in_h: " << in_h << ", in_w: " << in_w; int threads = 512; int group_size_x = (channel + 3) / 4; int group_size_y = batch; auto local_work_size = cl::NDRange{static_cast(threads), static_cast(1), static_cast(1)}; auto global_work_size = cl::NDRange{static_cast(group_size_x * threads), static_cast(group_size_y), static_cast(1)}; VLOG(4) << "local_work_size:[2D]:" << local_work_size[0] << " " << local_work_size[1] << " " << local_work_size[2]; VLOG(4) << "global_work_size:[2D]:" << global_work_size[0] << " " << global_work_size[1] << " " << global_work_size[2]; STL::stringstream kernel_key; kernel_key << kernel_func_name_ << build_options_; auto kernel = context.cl_context()->GetKernel(kernel_key.str()); auto* scale_img = scale_image_.data(); auto* bias_img = bias_image_.data(); float epsilon = instance_norm_param_->epsilon; int arg_idx = 0; cl_int status = kernel.setArg(arg_idx++, *x_img); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, *out_img); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, *scale_img); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, *bias_img); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, epsilon); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, in_h); CL_CHECK_FATAL(status); status = kernel.setArg(arg_idx++, in_w); CL_CHECK_FATAL(status); status = context.cl_context()->GetCommandQueue().enqueueNDRangeKernel( kernel, cl::NullRange, global_work_size, local_work_size, nullptr, event_.get()); CL_CHECK_FATAL(status); context.cl_wait_list()->emplace(out_img, event_); } protected: param_t* instance_norm_param_{nullptr}; std::string kernel_func_name_{"instance_norm"}; std::string build_options_{"-DCL_DTYPE_half"}; std::shared_ptr event_{new cl::Event}; Tensor scale_image_; Tensor bias_image_; }; } // namespace opencl } // namespace kernels } // namespace lite } // namespace paddle namespace ocl = paddle::lite::kernels::opencl; REGISTER_LITE_KERNEL(instance_norm, kOpenCL, kFP16, kImageDefault, ocl::InstanceNormImageCompute, ImageDefault) .BindInput("X", {LiteType::GetTensorTy(TARGET(kOpenCL), PRECISION(kFP16), DATALAYOUT(kImageDefault))}) .BindOutput("Y", {LiteType::GetTensorTy(TARGET(kOpenCL), PRECISION(kFP16), DATALAYOUT(kImageDefault))}) .BindInput("Scale", {LiteType::GetTensorTy(TARGET(kARM))}) .BindInput("Bias", {LiteType::GetTensorTy(TARGET(kARM))}) .BindOutput("SavedMean", {LiteType::GetTensorTy(TARGET(kARM))}) .BindOutput("SavedVariance", {LiteType::GetTensorTy(TARGET(kARM))}) .Finalize();