diff --git a/paddle/fluid/operators/elementwise/elementwise_pow_op.cc b/paddle/fluid/operators/elementwise/elementwise_pow_op.cc index 0273743c95add61f9d019a5a293336a80e1a33c0..b6389b97476683c3fb990fd08b258cb438a974c7 100644 --- a/paddle/fluid/operators/elementwise/elementwise_pow_op.cc +++ b/paddle/fluid/operators/elementwise/elementwise_pow_op.cc @@ -97,10 +97,9 @@ REGISTER_OPERATOR(elementwise_pow, ops::ElementwisePowOpMaker, ops::ElementwiseOpInferVarType, ops::ElementwisePowOpGradMaker, - ops::ElementwisePowOpGradMaker); -REGISTER_OPERATOR(elementwise_pow_grad, - ops::ElementwiseOpGrad, + ops::ElementwisePowOpGradMaker, ops::ElementwisePowCompositeGradOpMaker); +REGISTER_OPERATOR(elementwise_pow_grad, ops::ElementwiseOpGrad); REGISTER_OP_VERSION(elementwise_pow) .AddCheckpoint( diff --git a/paddle/fluid/prim/api/composite_backward/composite_backward_api.h b/paddle/fluid/prim/api/composite_backward/composite_backward_api.h index 0ac61cd4a3ea074c2032dee9915793915c1a839d..286d3cae8de5dfd3926609e1b551003a46ce47f5 100644 --- a/paddle/fluid/prim/api/composite_backward/composite_backward_api.h +++ b/paddle/fluid/prim/api/composite_backward/composite_backward_api.h @@ -398,7 +398,7 @@ void elementwise_pow_grad(const Tensor& x, // dy = lnx * x^y auto lnx = log(x); auto x_pow_y = elementwise_pow(x, y); - auto dy_res = lnx * x_pow_y; + auto dy_res = lnx * x_pow_y * out_grad; if (x.dims() != y.dims()) { // Maybe need reduce here phi::DDim reduce_dim = get_reduce_dims(y.dims(), x.dims()); @@ -418,7 +418,7 @@ void elementwise_pow_grad(const Tensor& x, // dx = y * x^(y-1) auto tmp_z = y - 1.0; auto x_pow_z = elementwise_pow(x, tmp_z); - auto dx_res = y * x_pow_z; + auto dx_res = y * x_pow_z * out_grad; if (y.dims() != x.dims()) { // Maybe need reduce here auto reduce_dim = get_reduce_dims(x.dims(), y.dims());