From 87843bebde86dc81a7c5504a4c482ae6796e3d94 Mon Sep 17 00:00:00 2001 From: hong19860320 <9973393+hong19860320@users.noreply.github.com> Date: Mon, 24 Aug 2020 10:20:30 +0800 Subject: [PATCH] Fix the sample code of SELU/selu (#26565) --- python/paddle/nn/functional/activation.py | 2 +- python/paddle/nn/layer/activation.py | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/python/paddle/nn/functional/activation.py b/python/paddle/nn/functional/activation.py index bd975106bdf..6acb806403e 100644 --- a/python/paddle/nn/functional/activation.py +++ b/python/paddle/nn/functional/activation.py @@ -665,7 +665,7 @@ def selu(x, paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) out = F.selu(x) # [[0, 1.050701],[2.101402, 3.152103]] """ if in_dygraph_mode(): diff --git a/python/paddle/nn/layer/activation.py b/python/paddle/nn/layer/activation.py index 6373b058832..bb0bd5f70f1 100644 --- a/python/paddle/nn/layer/activation.py +++ b/python/paddle/nn/layer/activation.py @@ -570,7 +570,7 @@ class SELU(layers.Layer): paddle.disable_static() - x = paddle.to_tensor(np.array([[0, 1],[2, 3]])) + x = paddle.to_tensor(np.array([[0.0, 1.0],[2.0, 3.0]])) m = paddle.nn.SELU() out = m(x) # [[0, 1.050701],[2.101402, 3.152103]] """ -- GitLab