diff --git a/python/paddle/nn/layer/activation.py b/python/paddle/nn/layer/activation.py index 02a1d297e83ea4f21b3f1a9cb85b950e5959dc08..1b82daedea8e923a52bd38819daf9a2ce6515479 100644 --- a/python/paddle/nn/layer/activation.py +++ b/python/paddle/nn/layer/activation.py @@ -210,9 +210,6 @@ class ReLU(layers.Layer): class LeakyReLU(layers.Layer): """ - :alias_main: paddle.nn.LeakyReLU - :alias: paddle.nn.LeakyReLU,paddle.nn.layer.LeakyReLU,paddle.nn.layer.activation.LeakyReLU - Leaky ReLU Activation. .. math: @@ -220,36 +217,35 @@ class LeakyReLU(layers.Layer): out = max(x, alpha * x) Parameters: - alpha (float, optional): Slope of the activation function at x < 0. Default: 0.01. - inplace (bool, optional): If inplace is True, the input and output of - ``LeakyReLU`` are the same variable. Otherwise, the input and output of - ``LeakyReLU`` are different variables. Default False. Note that if x is - more than one OPs' input, inplace must be False. Default: False. + alpha (float, optional): Slope of the activation function at :math:`x < 0` . + Default: 0.01. + name (str, optional): Name for the operation (optional, default is None). + For more information, please refer to :ref:`api_guide_Name`. - Returns: - None + Shape: + - input: Tensor with any shape. + - output: Tensor with the same shape as input. Examples: .. code-block:: python - import paddle.fluid as fluid - import paddle.nn as nn - import numpy as np + import paddle + import numpy as np - data = np.array([-2, 0, 1]).astype('float32') - lrelu = nn.LeakyReLU() - with fluid.dygraph.guard(): - data = fluid.dygraph.to_variable(data) - res = lrelu(data) # [-0.02, 0, 1] + paddle.enable_imperative() + + lrelu = paddle.nn.LeakyReLU() + x = paddle.imperative.to_variable(np.array([-2, 0, 1], 'float32')) + out = lrelu(x) # [-0.02, 0, 1] """ - def __init__(self, alpha=1e-2, inplace=False): + def __init__(self, alpha=1e-2, name=None): super(LeakyReLU, self).__init__() self._alpha = alpha - self._inplace = inplace + self._name = name - def forward(self, input): - return functional.leaky_relu(input, self._alpha, self._inplace) + def forward(self, x): + return functional.leaky_relu(x, self._alpha, self._name) class Sigmoid(layers.Layer):