From f1366d58326a78f6ac18c712e816c60074aa9a1d Mon Sep 17 00:00:00 2001 From: Chen Weihang Date: Fri, 31 Dec 2021 10:46:27 +0800 Subject: [PATCH] replace contextt to context (#38619) --- paddle/pten/kernels/complex_kernel.h | 4 ++-- paddle/pten/kernels/copy_kernel.h | 4 ++-- paddle/pten/kernels/cpu/copy_kernel.cc | 4 ++-- paddle/pten/kernels/cpu/dot_kernel.cc | 4 ++-- paddle/pten/kernels/dot_kernel.h | 4 ++-- paddle/pten/kernels/flatten_kernel.cc | 10 +++++----- paddle/pten/kernels/flatten_kernel.h | 8 ++++---- paddle/pten/kernels/full_kernel.h | 8 ++++---- paddle/pten/kernels/gpu/copy_kernel.cu | 4 ++-- paddle/pten/kernels/gpu/dot_kernel.cu | 4 ++-- paddle/pten/kernels/impl/complex_kernel_impl.h | 6 +++--- paddle/pten/kernels/impl/full_kernel_impl.h | 18 +++++++++--------- paddle/pten/kernels/impl/scale_kernel_impl.h | 4 ++-- paddle/pten/kernels/impl/sign_kernel_impl.h | 4 ++-- paddle/pten/kernels/reshape_kernel.cc | 8 ++++---- paddle/pten/kernels/reshape_kernel.h | 8 ++++---- paddle/pten/kernels/scale_kernel.h | 4 ++-- paddle/pten/kernels/sign_kernel.h | 4 ++-- paddle/pten/kernels/xpu/copy_kernel.cc | 4 ++-- 19 files changed, 57 insertions(+), 57 deletions(-) diff --git a/paddle/pten/kernels/complex_kernel.h b/paddle/pten/kernels/complex_kernel.h index b57e6d0fb4..dfe8fff43e 100644 --- a/paddle/pten/kernels/complex_kernel.h +++ b/paddle/pten/kernels/complex_kernel.h @@ -18,7 +18,7 @@ limitations under the License. */ namespace pten { -template -void Conj(const ContextT& dev_ctx, const DenseTensor& x, DenseTensor* out); +template +void Conj(const Context& dev_ctx, const DenseTensor& x, DenseTensor* out); } // namespace pten diff --git a/paddle/pten/kernels/copy_kernel.h b/paddle/pten/kernels/copy_kernel.h index d095d18a37..a481908892 100644 --- a/paddle/pten/kernels/copy_kernel.h +++ b/paddle/pten/kernels/copy_kernel.h @@ -18,8 +18,8 @@ limitations under the License. */ namespace pten { -template -void Copy(const ContextT& dev_ctx, +template +void Copy(const Context& dev_ctx, const DenseTensor& src, bool blocking, DenseTensor* dst); diff --git a/paddle/pten/kernels/cpu/copy_kernel.cc b/paddle/pten/kernels/cpu/copy_kernel.cc index 6a81579eb4..f3c4156fcd 100644 --- a/paddle/pten/kernels/cpu/copy_kernel.cc +++ b/paddle/pten/kernels/cpu/copy_kernel.cc @@ -25,8 +25,8 @@ limitations under the License. */ namespace pten { // NOTE(chenweihang): blocking is useless in cpu kernel -template -void Copy(const ContextT& dev_ctx, +template +void Copy(const Context& dev_ctx, const DenseTensor& src, bool blocking, DenseTensor* dst) { diff --git a/paddle/pten/kernels/cpu/dot_kernel.cc b/paddle/pten/kernels/cpu/dot_kernel.cc index c485cc8ac5..247ad1216a 100644 --- a/paddle/pten/kernels/cpu/dot_kernel.cc +++ b/paddle/pten/kernels/cpu/dot_kernel.cc @@ -22,8 +22,8 @@ namespace pten { -template -void Dot(const ContextT& dev_ctx, +template +void Dot(const Context& dev_ctx, const DenseTensor& x, const DenseTensor& y, DenseTensor* out) { diff --git a/paddle/pten/kernels/dot_kernel.h b/paddle/pten/kernels/dot_kernel.h index f6db41cbbe..9924749cd2 100644 --- a/paddle/pten/kernels/dot_kernel.h +++ b/paddle/pten/kernels/dot_kernel.h @@ -18,8 +18,8 @@ namespace pten { -template -void Dot(const ContextT& dev_ctx, +template +void Dot(const Context& dev_ctx, const DenseTensor& x, const DenseTensor& y, DenseTensor* out); diff --git a/paddle/pten/kernels/flatten_kernel.cc b/paddle/pten/kernels/flatten_kernel.cc index 9460574170..df8238cbf3 100644 --- a/paddle/pten/kernels/flatten_kernel.cc +++ b/paddle/pten/kernels/flatten_kernel.cc @@ -21,8 +21,8 @@ namespace pten { -template -void Flatten(const ContextT& dev_ctx, +template +void Flatten(const Context& dev_ctx, const DenseTensor& x, int start_axis, int stop_axis, @@ -35,14 +35,14 @@ void Flatten(const ContextT& dev_ctx, // TODO(yuanrisheng): this kernel is for training and xshape is a Intermediate // Output Tensor, // is there a more flexible way to deal with this case? -template -void FlattenWithXShape(const ContextT& dev_ctx, +template +void FlattenWithXShape(const Context& dev_ctx, const DenseTensor& x, int start_axis, int stop_axis, DenseTensor* out, DenseTensor* xshape) { - Flatten(dev_ctx, x, start_axis, stop_axis, out); + Flatten(dev_ctx, x, start_axis, stop_axis, out); funcs::SetXShape(x, xshape); } diff --git a/paddle/pten/kernels/flatten_kernel.h b/paddle/pten/kernels/flatten_kernel.h index 6ce0a2be20..5a0445489b 100644 --- a/paddle/pten/kernels/flatten_kernel.h +++ b/paddle/pten/kernels/flatten_kernel.h @@ -18,15 +18,15 @@ limitations under the License. */ namespace pten { -template -void Flatten(const ContextT& dev_ctx, +template +void Flatten(const Context& dev_ctx, const DenseTensor& x, int start_axis, int stop_axis, DenseTensor* out); -template -void FlattenWithXShape(const ContextT& dev_ctx, +template +void FlattenWithXShape(const Context& dev_ctx, const DenseTensor& x, int start_axis, int stop_axis, diff --git a/paddle/pten/kernels/full_kernel.h b/paddle/pten/kernels/full_kernel.h index d1139cf9ec..5bf6e37c36 100644 --- a/paddle/pten/kernels/full_kernel.h +++ b/paddle/pten/kernels/full_kernel.h @@ -20,13 +20,13 @@ namespace pten { -template -void Full(const ContextT& dev_ctx, +template +void Full(const Context& dev_ctx, const ScalarArray& shape, const Scalar& val, DenseTensor* out); -template -void FullLike(const ContextT& dev_ctx, const Scalar& val, DenseTensor* out); +template +void FullLike(const Context& dev_ctx, const Scalar& val, DenseTensor* out); } // namespace pten diff --git a/paddle/pten/kernels/gpu/copy_kernel.cu b/paddle/pten/kernels/gpu/copy_kernel.cu index cb9f8054d3..877a06ce33 100644 --- a/paddle/pten/kernels/gpu/copy_kernel.cu +++ b/paddle/pten/kernels/gpu/copy_kernel.cu @@ -24,8 +24,8 @@ limitations under the License. */ namespace pten { -template -void Copy(const ContextT& dev_ctx, +template +void Copy(const Context& dev_ctx, const DenseTensor& src, bool blocking, DenseTensor* dst) { diff --git a/paddle/pten/kernels/gpu/dot_kernel.cu b/paddle/pten/kernels/gpu/dot_kernel.cu index 7742e57a02..6b66d45b7d 100644 --- a/paddle/pten/kernels/gpu/dot_kernel.cu +++ b/paddle/pten/kernels/gpu/dot_kernel.cu @@ -24,8 +24,8 @@ namespace pten { -template -void Dot(const ContextT& dev_ctx, +template +void Dot(const Context& dev_ctx, const DenseTensor& x, const DenseTensor& y, DenseTensor* out) { diff --git a/paddle/pten/kernels/impl/complex_kernel_impl.h b/paddle/pten/kernels/impl/complex_kernel_impl.h index 7b5cabd680..6f3a6049fa 100644 --- a/paddle/pten/kernels/impl/complex_kernel_impl.h +++ b/paddle/pten/kernels/impl/complex_kernel_impl.h @@ -20,13 +20,13 @@ namespace pten { -template -void Conj(const ContextT& dev_ctx, const DenseTensor& x, DenseTensor* out) { +template +void Conj(const Context& dev_ctx, const DenseTensor& x, DenseTensor* out) { auto numel = x.numel(); auto* x_data = x.data(); auto* out_data = out->mutable_data(); - paddle::platform::ForRange for_range(dev_ctx, numel); + paddle::platform::ForRange for_range(dev_ctx, numel); paddle::operators::math::ConjFunctor functor(x_data, numel, out_data); for_range(functor); } diff --git a/paddle/pten/kernels/impl/full_kernel_impl.h b/paddle/pten/kernels/impl/full_kernel_impl.h index 7076bb51b3..c77b7a7077 100644 --- a/paddle/pten/kernels/impl/full_kernel_impl.h +++ b/paddle/pten/kernels/impl/full_kernel_impl.h @@ -24,24 +24,24 @@ limitations under the License. */ namespace pten { -template -void fill_(const DeviceContext& context, DenseTensor* tensor, VType val) { +template +void FullValue(const Context& dev_ctx, DenseTensor* tensor, VType val) { tensor->mutable_data(); auto t = pten::EigenVector::Flatten(*tensor); - t.device(*context.eigen_device()) = t.constant(static_cast(val)); + t.device(*dev_ctx.eigen_device()) = t.constant(static_cast(val)); } -template -void Full(const ContextT& dev_ctx, +template +void Full(const Context& dev_ctx, const ScalarArray& shape, const Scalar& val, DenseTensor* out) { out->Resize(paddle::framework::make_ddim(shape.GetData())); - fill_(dev_ctx, out, val.to()); + FullValue(dev_ctx, out, val.to()); } -template -void FullLike(const ContextT& dev_ctx, const Scalar& val, DenseTensor* out) { +template +void FullLike(const Context& dev_ctx, const Scalar& val, DenseTensor* out) { auto value = val.to(); using CommonType = typename std::common_type< float, @@ -66,7 +66,7 @@ void FullLike(const ContextT& dev_ctx, const Scalar& val, DenseTensor* out) { static_cast(std::numeric_limits::lowest()), static_cast(std::numeric_limits::max()), static_cast(value))); - fill_(dev_ctx, out, value); + FullValue(dev_ctx, out, value); } } // namespace pten diff --git a/paddle/pten/kernels/impl/scale_kernel_impl.h b/paddle/pten/kernels/impl/scale_kernel_impl.h index 421bb9f7b0..937b3115e6 100644 --- a/paddle/pten/kernels/impl/scale_kernel_impl.h +++ b/paddle/pten/kernels/impl/scale_kernel_impl.h @@ -23,8 +23,8 @@ limitations under the License. */ namespace pten { -template -void Scale(const ContextT& dev_ctx, +template +void Scale(const Context& dev_ctx, const DenseTensor& x, const Scalar& scale, float bias, diff --git a/paddle/pten/kernels/impl/sign_kernel_impl.h b/paddle/pten/kernels/impl/sign_kernel_impl.h index 088690ec64..d663808f03 100644 --- a/paddle/pten/kernels/impl/sign_kernel_impl.h +++ b/paddle/pten/kernels/impl/sign_kernel_impl.h @@ -22,8 +22,8 @@ limitations under the License. */ namespace pten { -template -void Sign(const ContextT& dev_ctx, const DenseTensor& x, DenseTensor* out) { +template +void Sign(const Context& dev_ctx, const DenseTensor& x, DenseTensor* out) { out->mutable_data(); auto eigen_out = pten::EigenVector::Flatten(*out); auto eigen_x = pten::EigenVector::Flatten(x); diff --git a/paddle/pten/kernels/reshape_kernel.cc b/paddle/pten/kernels/reshape_kernel.cc index ea1f03a017..0535ea20c8 100644 --- a/paddle/pten/kernels/reshape_kernel.cc +++ b/paddle/pten/kernels/reshape_kernel.cc @@ -21,8 +21,8 @@ namespace pten { -template -void Reshape(const ContextT& dev_ctx, +template +void Reshape(const Context& dev_ctx, const DenseTensor& x, const ScalarArray& shape, DenseTensor* out) { @@ -36,8 +36,8 @@ void Reshape(const ContextT& dev_ctx, out->ResetLoD(x.lod()); } -template -void ReshapeWithXShape(const ContextT& dev_ctx, +template +void ReshapeWithXShape(const Context& dev_ctx, const DenseTensor& x, const ScalarArray& shape, DenseTensor* xshape, diff --git a/paddle/pten/kernels/reshape_kernel.h b/paddle/pten/kernels/reshape_kernel.h index d9ccd0449b..b10e31a434 100644 --- a/paddle/pten/kernels/reshape_kernel.h +++ b/paddle/pten/kernels/reshape_kernel.h @@ -19,14 +19,14 @@ limitations under the License. */ namespace pten { -template -void Reshape(const ContextT& dev_ctx, +template +void Reshape(const Context& dev_ctx, const DenseTensor& x, const ScalarArray& shape, DenseTensor* out); -template -void ReshapeWithXShape(const ContextT& dev_ctx, +template +void ReshapeWithXShape(const Context& dev_ctx, const DenseTensor& x, const ScalarArray& shape, DenseTensor* xshape, diff --git a/paddle/pten/kernels/scale_kernel.h b/paddle/pten/kernels/scale_kernel.h index bb3c1968fc..5908050029 100644 --- a/paddle/pten/kernels/scale_kernel.h +++ b/paddle/pten/kernels/scale_kernel.h @@ -19,8 +19,8 @@ limitations under the License. */ namespace pten { -template -void Scale(const ContextT& dev_ctx, +template +void Scale(const Context& dev_ctx, const DenseTensor& x, const Scalar& scale, float bias, diff --git a/paddle/pten/kernels/sign_kernel.h b/paddle/pten/kernels/sign_kernel.h index bbb3f45c9a..2cf5ca973f 100644 --- a/paddle/pten/kernels/sign_kernel.h +++ b/paddle/pten/kernels/sign_kernel.h @@ -18,7 +18,7 @@ limitations under the License. */ namespace pten { -template -void Sign(const ContextT& dev_ctx, const DenseTensor& x, DenseTensor* out); +template +void Sign(const Context& dev_ctx, const DenseTensor& x, DenseTensor* out); } // namespace pten diff --git a/paddle/pten/kernels/xpu/copy_kernel.cc b/paddle/pten/kernels/xpu/copy_kernel.cc index 479ef50836..190eb39e22 100644 --- a/paddle/pten/kernels/xpu/copy_kernel.cc +++ b/paddle/pten/kernels/xpu/copy_kernel.cc @@ -24,8 +24,8 @@ limitations under the License. */ namespace pten { -template -void Copy(const ContextT& dev_ctx, +template +void Copy(const Context& dev_ctx, const DenseTensor& src, bool blocking, DenseTensor* dst) { -- GitLab