diff --git a/paddle/fluid/operators/elementwise/elementwise_max_op.cc b/paddle/fluid/operators/elementwise/elementwise_max_op.cc index cc27bab72005752af372ee1a273258033e028c18..91da732ef0d3dfbda5d9b7734071ec5831bcfa3f 100644 --- a/paddle/fluid/operators/elementwise/elementwise_max_op.cc +++ b/paddle/fluid/operators/elementwise/elementwise_max_op.cc @@ -124,13 +124,17 @@ REGISTER_OP_CPU_KERNEL( ops::ElementwiseMaxKernel, ops::ElementwiseMaxKernel, ops::ElementwiseMaxKernel, - ops::ElementwiseMaxKernel); + ops::ElementwiseMaxKernel, + ops::ElementwiseMaxKernel); REGISTER_OP_CPU_KERNEL( elementwise_max_grad, ops::ElementwiseMaxGradKernel, ops::ElementwiseMaxGradKernel, ops::ElementwiseMaxGradKernel, - ops::ElementwiseMaxGradKernel); + ops::ElementwiseMaxGradKernel, + ops::ElementwiseMaxGradKernel); REGISTER_OP_VERSION(elementwise_max) .AddCheckpoint( diff --git a/paddle/fluid/operators/elementwise/elementwise_max_op.cu b/paddle/fluid/operators/elementwise/elementwise_max_op.cu index 7433c505f472a28c7be118856ee8fa410025449a..123332a4a23de5c9534c8523993b87d8738f9869 100644 --- a/paddle/fluid/operators/elementwise/elementwise_max_op.cu +++ b/paddle/fluid/operators/elementwise/elementwise_max_op.cu @@ -69,6 +69,8 @@ REGISTER_OP_CUDA_KERNEL( elementwise_max, ops::ElementwiseMaxKernel, + ops::ElementwiseMaxKernel, ops::ElementwiseMaxKernel, ops::ElementwiseMaxKernel, ops::ElementwiseMaxKernel, @@ -77,6 +79,8 @@ REGISTER_OP_CUDA_KERNEL( elementwise_max_grad, ops::ElementwiseMaxGradKernel, + ops::ElementwiseMaxGradKernel, ops::ElementwiseMaxGradKernel, ops::ElementwiseMaxGradKernel, ops::ElementwiseMaxGradKernel, diff --git a/python/paddle/fluid/tests/unittests/op_test.py b/python/paddle/fluid/tests/unittests/op_test.py index 85423df3d382831738c2c64ea845d0661f9cdbb7..ec3b68086b06593b035b60825a52b0ec32b8281d 100644 --- a/python/paddle/fluid/tests/unittests/op_test.py +++ b/python/paddle/fluid/tests/unittests/op_test.py @@ -168,8 +168,10 @@ def get_numeric_gradient(place, elif tensor_to_check._dtype() == core.VarDesc.VarType.BF16: numpy_tensor = np.array(tensor).astype(np.uint16) numpy_tensor = numpy_tensor.flatten() - return struct.unpack('