未验证 提交 12473236 编写于 作者: Z zhangyingying520 提交者: GitHub

修复paddle.nn.ChannelShuffle等 API 的文档 (#48742)

* 1

* Update vision.py

; test=docs_preview

* Update vision.py

* Update loss.py

修改缩进; test=docs_preview

* Update loss.py

* test=document_fix

* test=document_fix

* for codestyle; test=document_fix
Co-authored-by: NLigoml <39876205+Ligoml@users.noreply.github.com>
Co-authored-by: 梦柳's avatarLigoml <limengliu@tiaozhan.com>
上级 b89cea33
...@@ -2346,21 +2346,21 @@ def cross_entropy( ...@@ -2346,21 +2346,21 @@ def cross_entropy(
): ):
r""" r"""
By default, this operator implements the cross entropy loss function with softmax. This function By default, the cross entropy loss function is implemented using softmax. This function
combines the calculation of the softmax operation and the cross entropy loss function combines the calculation of the softmax operation and the cross entropy loss function
to provide a more numerically stable computing. to provide a more numerically stable computing.
This operator will calculate the cross entropy loss function without softmax when use_softmax=False. Calculate the cross entropy loss function without softmax when use_softmax=False.
By default, this operator will calculate the mean of the result, and you can also affect By default, calculate the mean of the result, and you can also affect
the default behavior by using the reduction parameter. Please refer to the part of the default behavior by using the reduction parameter. Please refer to the part of
parameters for details. parameters for details.
This operator can be used to calculate the softmax cross entropy loss with soft and hard labels. Can be used to calculate the softmax cross entropy loss with soft and hard labels.
Where, the hard labels mean the actual label value, 0, 1, 2, etc. And the soft labels Where, the hard labels mean the actual label value, 0, 1, 2, etc. And the soft labels
mean the probability of the actual label, 0.6, 0.8, 0.2, etc. mean the probability of the actual label, 0.6, 0.8, 0.2, etc.
The calculation of this operator includes the following two steps. The calculation includes the following two steps.
- **1.softmax cross entropy** - **1.softmax cross entropy**
...@@ -3480,7 +3480,7 @@ def cosine_embedding_loss( ...@@ -3480,7 +3480,7 @@ def cosine_embedding_loss(
input1, input2, label, margin=0, reduction='mean', name=None input1, input2, label, margin=0, reduction='mean', name=None
): ):
r""" r"""
This operator computes the cosine embedding loss of Tensor ``input1``, ``input2`` and ``label`` as follows. Compute the cosine embedding loss of Tensor ``input1``, ``input2`` and ``label`` as follows.
If label = 1, then the loss value can be calculated as follow: If label = 1, then the loss value can be calculated as follow:
...@@ -3497,11 +3497,11 @@ def cosine_embedding_loss( ...@@ -3497,11 +3497,11 @@ def cosine_embedding_loss(
cos(x1, x2) = \frac{x1 \cdot{} x2}{\Vert x1 \Vert_2 * \Vert x2 \Vert_2} cos(x1, x2) = \frac{x1 \cdot{} x2}{\Vert x1 \Vert_2 * \Vert x2 \Vert_2}
Parameters: Parameters:
input1 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, 'M' means the length of input array. input1 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, which can be 0, 'M' means the length of input array.
Available dtypes are float32, float64. Available dtypes are float32, float64.
input2 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, 'M' means the length of input array. input2 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, which can be 0, 'M' means the length of input array.
Available dtypes are float32, float64. Available dtypes are float32, float64.
label (Tensor): tensor with shape: [N] or [1]. The target labels values should be -1 or 1. label (Tensor): tensor with shape: [N] or [1], 'N' means the length of input array. The target labels values should be -1 or 1.
Available dtypes are int32, int64, float32, float64. Available dtypes are int32, int64, float32, float64.
margin (float, optional): Should be a number from :math:`-1` to :math:`1`, margin (float, optional): Should be a number from :math:`-1` to :math:`1`,
:math:`0` to :math:`0.5` is suggested. If :attr:`margin` is missing, the :math:`0` to :math:`0.5` is suggested. If :attr:`margin` is missing, the
......
...@@ -465,12 +465,12 @@ def pixel_unshuffle(x, downscale_factor, data_format="NCHW", name=None): ...@@ -465,12 +465,12 @@ def pixel_unshuffle(x, downscale_factor, data_format="NCHW", name=None):
def channel_shuffle(x, groups, data_format="NCHW", name=None): def channel_shuffle(x, groups, data_format="NCHW", name=None):
""" """
This API implements channel shuffle operation. This API implements channel shuffle operation.
See more details in :ref:`api_nn_vision_ChannelShuffle` . See more details in :ref:`api_nn_vision_ChannelShuffle`.
Parameters: Parameters:
x (Tensor): 4-D tensor, the data type should be float32 or float64. x (Tensor): 4-D tensor, the data type should be float32 or float64.
groups (int): Number of groups to divide channels in. groups (int): Number of groups to divide channels in.
data_format (str): The data format of the input and output data. An optional string of NCHW or NHWC. The default is NCHW. When it is NCHW, the data is stored in the order of [batch_size, input_channels, input_height, input_width]. data_format (str, optional): The data format of the input and output data. An optional string of NCHW or NHWC. The default is NCHW. When it is NCHW, the data is stored in the order of [batch_size, input_channels, input_height, input_width].
name (str, optional): Name for the operation (optional, default is None). Normally there is no need for user to set this property. For more information, please refer to :ref:`api_guide_Name`. name (str, optional): Name for the operation (optional, default is None). Normally there is no need for user to set this property. For more information, please refer to :ref:`api_guide_Name`.
Returns: Returns:
......
...@@ -138,21 +138,21 @@ class BCEWithLogitsLoss(Layer): ...@@ -138,21 +138,21 @@ class BCEWithLogitsLoss(Layer):
class CrossEntropyLoss(Layer): class CrossEntropyLoss(Layer):
r""" r"""
By default, this operator implements the cross entropy loss function with softmax. This function By default, the cross entropy loss function is implemented using softmax. This function
combines the calculation of the softmax operation and the cross entropy loss function combines the calculation of the softmax operation and the cross entropy loss function
to provide a more numerically stable computing. to provide a more numerically stable computing.
This operator will calculate the cross entropy loss function without softmax when use_softmax=False. Calculate the cross entropy loss function without softmax when use_softmax=False.
By default, this operator will calculate the mean of the result, and you can also affect By default, calculate the mean of the result, and you can also affect
the default behavior by using the reduction parameter. Please refer to the part of the default behavior by using the reduction parameter. Please refer to the part of
parameters for details. parameters for details.
This operator can be used to calculate the softmax cross entropy loss with soft and hard labels. Can be used to calculate the softmax cross entropy loss with soft and hard labels.
Where, the hard labels mean the actual label value, 0, 1, 2, etc. And the soft labels Where, the hard labels mean the actual label value, 0, 1, 2, etc. And the soft labels
mean the probability of the actual label, 0.6, 0.8, 0.2, etc. mean the probability of the actual label, 0.6, 0.8, 0.2, etc.
The calculation of this operator includes the following two steps. The calculation includes the following two steps.
- **I.softmax cross entropy** - **I.softmax cross entropy**
...@@ -277,8 +277,8 @@ class CrossEntropyLoss(Layer): ...@@ -277,8 +277,8 @@ class CrossEntropyLoss(Layer):
Shape: Shape:
- **input** (Tensor), the data type is float32, float64. Shape is - **input** (Tensor), the data type is float32, float64. Shape is :math:`[N_1, N_2, ..., N_k, C]`, where C is number of classes, ``k >= 1`` .
:math:`[N_1, N_2, ..., N_k, C]`, where C is number of classes , ``k >= 1`` .
Note: Note:
1. when use_softmax=True, it expects unscaled logits. This operator should not be used with the 1. when use_softmax=True, it expects unscaled logits. This operator should not be used with the
...@@ -1413,11 +1413,11 @@ class CosineEmbeddingLoss(Layer): ...@@ -1413,11 +1413,11 @@ class CosineEmbeddingLoss(Layer):
For more information, please refer to :ref:`api_guide_Name`. For more information, please refer to :ref:`api_guide_Name`.
Shape: Shape:
input1 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, 'M' means the length of input array. input1 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, which can be 0, 'M' means the length of input array.
Available dtypes are float32, float64. Available dtypes are float32, float64.
input2 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, 'M' means the length of input array. input2 (Tensor): tensor with shape: [N, M] or [M], 'N' means batch size, which can be 0, 'M' means the length of input array.
Available dtypes are float32, float64. Available dtypes are float32, float64.
label (Tensor): tensor with shape: [N] or [1]. The target labels values should be -1 or 1. label (Tensor): tensor with shape: [N] or [1], 'N' means the length of input array. The target labels values should be -1 or 1.
Available dtypes are int32, int64, float32, float64. Available dtypes are int32, int64, float32, float64.
output (Tensor): Tensor, the cosine embedding Loss of Tensor ``input1`` ``input2`` and ``label``. output (Tensor): Tensor, the cosine embedding Loss of Tensor ``input1`` ``input2`` and ``label``.
If `reduction` is ``'none'``, the shape of output loss is [N], the same as ``input`` . If `reduction` is ``'none'``, the shape of output loss is [N], the same as ``input`` .
......
...@@ -156,7 +156,7 @@ class PixelUnshuffle(Layer): ...@@ -156,7 +156,7 @@ class PixelUnshuffle(Layer):
class ChannelShuffle(Layer): class ChannelShuffle(Layer):
""" """
This operator divides channels in a tensor of shape [N, C, H, W] or [N, H, W, C] into g groups, Can divide channels in a tensor of shape [N, C, H, W] or [N, H, W, C] into g groups,
getting a tensor with the shape of [N, g, C/g, H, W] or [N, H, W, g, C/g], and transposes them getting a tensor with the shape of [N, g, C/g, H, W] or [N, H, W, g, C/g], and transposes them
as [N, C/g, g, H, W] or [N, H, W, g, C/g], then rearranges them to original tensor shape. This as [N, C/g, g, H, W] or [N, H, W, g, C/g], then rearranges them to original tensor shape. This
operation can improve the interaction between channels, using features efficiently. Please operation can improve the interaction between channels, using features efficiently. Please
...@@ -166,7 +166,7 @@ class ChannelShuffle(Layer): ...@@ -166,7 +166,7 @@ class ChannelShuffle(Layer):
Parameters: Parameters:
groups (int): Number of groups to divide channels in. groups (int): Number of groups to divide channels in.
data_format (str): The data format of the input and output data. An optional string of NCHW or NHWC. The default is NCHW. When it is NCHW, the data is stored in the order of [batch_size, input_channels, input_height, input_width]. data_format (str, optional): The data format of the input and output data. An optional string of NCHW or NHWC. The default is NCHW. When it is NCHW, the data is stored in the order of [batch_size, input_channels, input_height, input_width].
name (str, optional): Name for the operation (optional, default is None). Normally there is no need for user to set this property. For more information, please refer to :ref:`api_guide_Name`. name (str, optional): Name for the operation (optional, default is None). Normally there is no need for user to set this property. For more information, please refer to :ref:`api_guide_Name`.
Shape: Shape:
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册