diff --git a/python/paddle/nn/functional/activation.py b/python/paddle/nn/functional/activation.py index bd975106bdf1db0fac8cbbf0c22913c8016ea680..6acb806403ec782e664b9c173abbc29537fea3eb 100644 --- a/python/paddle/nn/functional/activation.py +++ b/python/paddle/nn/functional/activation.py @@ -665,7 +665,7 @@ def selu(x, paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) out = F.selu(x) # [[0, 1.050701],[2.101402, 3.152103]] """ if in_dygraph_mode(): diff --git a/python/paddle/nn/layer/activation.py b/python/paddle/nn/layer/activation.py index 6373b058832aab8da26d5257fd1aa8bf3a856155..bb0bd5f70f1f9006285a3fb200537d35b4cf6c30 100644 --- a/python/paddle/nn/layer/activation.py +++ b/python/paddle/nn/layer/activation.py @@ -570,7 +570,7 @@ class SELU(layers.Layer): paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) m = paddle.nn.SELU() out = m(x) # [[0, 1.050701],[2.101402, 3.152103]] """