// Copyright (c) 2019 PaddlePaddle Authors. All Rights Reserved.//// Licensed under the Apache License, Version 2.0 (the "License");// you may not use this file except in compliance with the License.// You may obtain a copy of the License at//// http://www.apache.org/licenses/LICENSE-2.0//// Unless required by applicable law or agreed to in writing, software// distributed under the License is distributed on an "AS IS" BASIS,// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.// See the License for the specific language governing permissions and// limitations under the License.#include "lite/kernels/x86/attention_padding_mask_compute.h"REGISTER_LITE_KERNEL(attention_padding_mask,kX86,kFloat,kNCHW,paddle::lite::kernels::x86::AttentionPaddingMaskCompute<float>,def).BindInput("X",{LiteType::GetTensorTy(TARGET(kX86))}).BindInput("Y",{LiteType::GetTensorTy(TARGET(kX86))}).BindOutput("out",{LiteType::GetTensorTy(TARGET(kX86))}).BindOutput("pad_begin",{LiteType::GetTensorTy(TARGET(kX86))}).Finalize();