diff --git a/mindspore/nn/layer/conv.py b/mindspore/nn/layer/conv.py index 0f457fce9f778d048d5d1aece59aa3bb9d991dd6..790bbe8e7236044b2723c1ebc12825c204b727f0 100644 --- a/mindspore/nn/layer/conv.py +++ b/mindspore/nn/layer/conv.py @@ -111,14 +111,14 @@ class Conv2d(_Conv): 2D convolution layer. Applies a 2D convolution over an input tensor which is typically of shape :math:`(N, C_{in}, H_{in}, W_{in})`, - where :math:`N` is batch size and :math:`C_{in}` is channel number. For each batch of shape - :math:`(C_{in}, H_{in}, W_{in})`, the formula is defined as: + where :math:`N` is batch size, :math:`C_{in}` is channel number, and :math:`H_{in}, W_{in})` are height and width. + For each batch of shape :math:`(C_{in}, H_{in}, W_{in})`, the formula is defined as: .. math:: out_j = \sum_{i=0}^{C_{in} - 1} ccor(W_{ij}, X_i) + b_j, - where :math:`ccor` is the cross correlation operator, :math:`C_{in}` is the input channel number, :math:`j` ranges + where :math:`ccor` is the cross-correlation operator, :math:`C_{in}` is the input channel number, :math:`j` ranges from :math:`0` to :math:`C_{out} - 1`, :math:`W_{ij}` corresponds to the :math:`i`-th channel of the :math:`j`-th filter and :math:`out_{j}` corresponds to the :math:`j`-th channel of the output. :math:`W_{ij}` is a slice of kernel and it has shape :math:`(\text{ks_h}, \text{ks_w})`, where :math:`\text{ks_h}` and @@ -162,8 +162,8 @@ class Conv2d(_Conv): Tensor borders. `padding` should be greater than or equal to 0. padding (Union[int, tuple[int]]): Implicit paddings on both sides of the input. If `padding` is one integer, - the padding of top, bottom, left and right is the same, equal to padding. If `padding` is a tuple - with four integers, the padding of top, bottom, left and right will be equal to padding[0], + the paddings of top, bottom, left and right are the same, equal to padding. If `padding` is a tuple + with four integers, the paddings of top, bottom, left and right will be equal to padding[0], padding[1], padding[2], and padding[3] accordingly. Default: 0. dilation (Union[int, tuple[int]]): The data type is int or a tuple of 2 integers. Specifies the dilation rate to use for dilated convolution. If set to be :math:`k > 1`, there will @@ -472,8 +472,8 @@ class Conv2dTranspose(_Conv): - valid: Adopted the way of discarding. padding (Union[int, tuple[int]]): Implicit paddings on both sides of the input. If `padding` is one integer, - the padding of top, bottom, left and right is the same, equal to padding. If `padding` is a tuple - with four integers, the padding of top, bottom, left and right will be equal to padding[0], + the paddings of top, bottom, left and right are the same, equal to padding. If `padding` is a tuple + with four integers, the paddings of top, bottom, left and right will be equal to padding[0], padding[1], padding[2], and padding[3] accordingly. Default: 0. dilation (Union[int, tuple[int]]): The data type is int or a tuple of 2 integers. Specifies the dilation rate to use for dilated convolution. If set to be :math:`k > 1`, there will @@ -856,8 +856,8 @@ class DepthwiseConv2d(Cell): Tensor borders. `padding` should be greater than or equal to 0. padding (Union[int, tuple[int]]): Implicit paddings on both sides of the input. If `padding` is one integer, - the padding of top, bottom, left and right is the same, equal to padding. If `padding` is a tuple - with four integers, the padding of top, bottom, left and right will be equal to padding[0], + the paddings of top, bottom, left and right are the same, equal to padding. If `padding` is a tuple + with four integers, the paddings of top, bottom, left and right will be equal to padding[0], padding[1], padding[2], and padding[3] accordingly. Default: 0. dilation (Union[int, tuple[int]]): The data type is int or a tuple of 2 integers. Specifies the dilation rate to use for dilated convolution. If set to be :math:`k > 1`, there will diff --git a/mindspore/ops/operations/_grad_ops.py b/mindspore/ops/operations/_grad_ops.py index c16ef4872aa0adc99474ef892509f0d60016caa5..8df5c55f7711de631d2429920697c822eabe4c44 100644 --- a/mindspore/ops/operations/_grad_ops.py +++ b/mindspore/ops/operations/_grad_ops.py @@ -284,12 +284,12 @@ class Conv2DBackpropFilter(PrimitiveWithInfer): Args: out_channel (int): The dimensionality of the output space. kernel_size (Union[int, tuple[int]]): The size of the convolution window. - pad_mode (str): "valid", "same", "pad" the mode to fill padding. Default: "valid". - pad (int): The pad value to fill. Default: 0. - mode (int): 0 Math convolutiuon, 1 cross-correlation convolution , + pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid". + pad (int): The pad value to be filled. Default: 0. + mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution , 2 deconvolution, 3 depthwise convolution. Default: 1. - stride (tuple): The stride to apply conv filter. Default: (1, 1). - dilation (tuple): Specifies the dilation rate to use for dilated convolution. Default: (1, 1, 1, 1). + stride (tuple): The stride to be applied to the convolution filter. Default: (1, 1). + dilation (tuple): Specifies the dilation rate to be used for the dilated convolution. Default: (1, 1, 1, 1). group (int): Splits input into groups. Default: 1. Returns: @@ -349,12 +349,12 @@ class DepthwiseConv2dNativeBackpropFilter(PrimitiveWithInfer): Args: channel_multiplier (int): The multipiler for the original output conv. kernel_size (int or tuple): The size of the conv kernel. - mode (int): 0 Math convolutiuon, 1 cross-correlation convolution, + mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution, 2 deconvolution,3 depthwise convolution. Defaul: 3. pad_mode (str): The mode to fill padding which can be: "valid", "same" or "pad". Default: "valid". - pad (int): The pad value to fill. Default: 0. + pad (int): The pad value to be filled. Default: 0. pads (tuple): The pad list like (top, bottom, left, right). Default: (0, 0, 0, 0). - stride (int): The stride to apply conv filter. Default: 1. + stride (int): The stride to be applied to the convolution filter. Default: 1. dilation (int): Specifies the space to use between kernel elements. Default: 1. group (int): Splits input into groups. Default: 1. @@ -410,12 +410,12 @@ class DepthwiseConv2dNativeBackpropInput(PrimitiveWithInfer): Args: channel_multiplier (int): The multipiler for the original output conv. kernel_size (int or tuple): The size of the conv kernel. - mode (int): 0 Math convolutiuon, 1 cross-correlation convolution , + mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution , 2 deconvolution,3 depthwise convolution. Default: 3. - pad_mode (str): "valid", "same", "pad" the mode to fill padding. Default: "valid". - pad (int): the pad value to fill. Default: 0. + pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid". + pad (int): The pad value to be filled. Default: 0. pads (tuple): The pad list like (top, bottom, left, right). Default: (0, 0, 0, 0). - stride (int): the stride to apply conv filter. Default: 1. + stride (int): The stride to be applied to the convolution filter. Default: 1. dilation (int): Specifies the space to use between kernel elements. Default: 1. group (int): Splits input into groups. Default: 1. diff --git a/mindspore/ops/operations/array_ops.py b/mindspore/ops/operations/array_ops.py index c42c2505520072662849346efd05e94532b20fea..8e6fd299f1d753689cd8d1a078fb15021e397ffd 100644 --- a/mindspore/ops/operations/array_ops.py +++ b/mindspore/ops/operations/array_ops.py @@ -292,7 +292,7 @@ class IsSubClass(PrimitiveWithInfer): Check whether one type is sub class of another type. Inputs: - - **sub_type** (mindspore.dtype) - The type to be check. Only constant value is allowed. + - **sub_type** (mindspore.dtype) - The type to be checked. Only constant value is allowed. - **type_** (mindspore.dtype) - The target type. Only constant value is allowed. Outputs: @@ -326,7 +326,7 @@ class IsInstance(PrimitiveWithInfer): Check whether an object is an instance of a target type. Inputs: - - **inst** (Any Object) - The instance to be check. Only constant value is allowed. + - **inst** (Any Object) - The instance to be checked. Only constant value is allowed. - **type_** (mindspore.dtype) - The target type. Only constant value is allowed. Outputs: @@ -1100,7 +1100,7 @@ class InvertPermutation(PrimitiveWithInfer): Only constant value is allowed. Outputs: - tuple[int]. the lenth is same as input. + tuple[int]. It has the same length as the input. Examples: >>> invert = P.InvertPermutation() @@ -2355,15 +2355,15 @@ class DiagPart(PrimitiveWithInfer): class Eye(PrimitiveWithInfer): """ - Creates a tensor with ones on the diagonal and zeros elsewhere. + Creates a tensor with ones on the diagonal and zeros the rest. Inputs: - - **n** (int) - Number of rows of returned tensor - - **m** (int) - Number of columns of returned tensor + - **n** (int) - The number of rows of returned tensor + - **m** (int) - The number of columns of returned tensor - **t** (mindspore.dtype) - MindSpore's dtype, The data type of the returned tensor. Outputs: - Tensor, a tensor with ones on the diagonal and zeros elsewhere. + Tensor, a tensor with ones on the diagonal and the rest of elements are zero. Examples: >>> eye = P.Eye() @@ -3453,8 +3453,8 @@ class InplaceUpdate(PrimitiveWithInfer): Inputs: - **x** (Tensor) - A tensor which to be inplace updated. It can be one of the following data types: - float32, float16, int32. - - **v** (Tensor) - A tensor of the same type as `x`. Same dimension size as `x` except + float32, float16 and int32. + - **v** (Tensor) - A tensor with the same type as `x` and the same dimension size as `x` except the first dimension, which must be the same as the size of `indices`. Outputs: diff --git a/mindspore/ops/operations/control_ops.py b/mindspore/ops/operations/control_ops.py index 7ef6da2e913ceac10348ec1bdcbafed55a911355..a7057549c4731eb9efd26ad8655bcb25a36838f6 100644 --- a/mindspore/ops/operations/control_ops.py +++ b/mindspore/ops/operations/control_ops.py @@ -26,23 +26,24 @@ class ControlDepend(Primitive): Adds control dependency relation between source and destination operation. In many cases, we need to control the execution order of operations. ControlDepend is designed for this. - ControlDepend will indicate the execution engine to run the operations in specific order. ControlDepend + ControlDepend will instruct the execution engine to run the operations in a specific order. ControlDepend tells the engine that the destination operations should depend on the source operation which means the source operations should be executed before the destination. Note: This operation does not work in `PYNATIVE_MODE`. Args: - depend_mode (int): Use 0 for normal depend, 1 for depend on operations that used the parameter. Default: 0. + depend_mode (int): Use 0 for a normal dependency relation. Use 1 to depends on operations which using Parameter + as its input. Default: 0. Inputs: - **src** (Any) - The source input. It can be a tuple of operations output or a single operation output. We do not concern about the input data, but concern about the operation that generates the input data. - If `depend_mode = 1` is specified and the source input is parameter, we will try to find the operations that + If `depend_mode` is 1 and the source input is Parameter, we will try to find the operations that used the parameter as input. - **dst** (Any) - The destination input. It can be a tuple of operations output or a single operation output. We do not concern about the input data, but concern about the operation that generates the input data. - If `depend_mode = 1` is specified and the source input is parameter, we will try to find the operations that + If `depend_mode` is 1 and the source input is Parameter, we will try to find the operations that used the parameter as input. Outputs: @@ -80,7 +81,7 @@ class GeSwitch(PrimitiveWithInfer): """ Adds control switch to data. - Switch data to flow into false or true branch depend on the condition. If the condition is true, + Switch data flows into false or true branch depending on the condition. If the condition is true, the true branch will be activated, or vise verse. Inputs: diff --git a/mindspore/ops/operations/debug_ops.py b/mindspore/ops/operations/debug_ops.py index 4123122ea21c8d3426af98cad7311a68f6e727a1..57357f4b25680562e9c82f29d4095e7eca214185 100644 --- a/mindspore/ops/operations/debug_ops.py +++ b/mindspore/ops/operations/debug_ops.py @@ -248,14 +248,14 @@ class InsertGradientOf(PrimitiveWithInfer): class HookBackward(PrimitiveWithInfer): """ - Used as tag to hook gradient in intermediate variables. Note that this function + This operation is used as a tag to hook gradient in intermediate variables. Note that this function is only supported in Pynative Mode. Note: The hook function should be defined like `hook_fn(grad) -> Tensor or None`, - which grad is the gradient passed to the primitive and gradient may be - modified and passed to nex primitive. the difference between hook function and - callback of InsertGradientOf is that hook function is executed in python + where grad is the gradient passed to the primitive and gradient may be + modified and passed to next primitive. The difference between a hook function and + callback of InsertGradientOf is that a hook function is executed in the python environment while callback will be parsed and added to the graph. Args: diff --git a/mindspore/ops/operations/image_ops.py b/mindspore/ops/operations/image_ops.py index b24b6d564e37a4fc94c33e902949f351cde2abf1..66435c39fbfa003b727f68af559c12e76479fa65 100644 --- a/mindspore/ops/operations/image_ops.py +++ b/mindspore/ops/operations/image_ops.py @@ -29,9 +29,9 @@ class CropAndResize(PrimitiveWithInfer): In case that the output shape depends on crop_size, the crop_size should be constant. Args: - method (str): An optional string specifying the sampling method for resizing. - It can be either "bilinear" or "nearest" and default to "bilinear" - extrapolation_value (float): An optional float defaults to 0. Value used for extrapolation, when applicable. + method (str): An optional string that specifies the sampling method for resizing. + It can be either "bilinear" or "nearest". Default: "bilinear" + extrapolation_value (float): An optional float value used extrapolation, if applicable. Default: 0. Inputs: - **x** (Tensor) - The input image must be a 4-D tensor of shape [batch, image_height, image_width, depth]. diff --git a/mindspore/ops/operations/math_ops.py b/mindspore/ops/operations/math_ops.py index 8355e4aa29b5d00ca6132f20104745d7c0e4d117..798716996d087afe8fa7ec9a324befd041ca8b57 100644 --- a/mindspore/ops/operations/math_ops.py +++ b/mindspore/ops/operations/math_ops.py @@ -122,7 +122,7 @@ class TensorAdd(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -957,11 +957,11 @@ class InplaceAdd(PrimitiveWithInfer): Args: indices (Union[int, tuple]): Indices into the left-most dimension of x, and determines which rows of x - to add with v. It is a int or tuple, whose value is in [0, the first dimension size of x). + to add with v. It is an integer or a tuple, whose value is in [0, the first dimension size of x). Inputs: - **input_x** (Tensor) - The first input is a tensor whose data type is float16, float32 or int32. - - **input_v** (Tensor) - The second input is a tensor who has the same dimension sizes as x except + - **input_v** (Tensor) - The second input is a tensor that has the same dimension sizes as x except the first dimension, which must be the same as indices's size. It has the same data type with `input_x`. Outputs: @@ -1015,7 +1015,7 @@ class InplaceSub(PrimitiveWithInfer): Args: indices (Union[int, tuple]): Indices into the left-most dimension of x, and determines which rows of x - to sub with v. It is a int or tuple, whose value is in [0, the first dimension size of x). + to subtract with v. It is a int or tuple, whose value is in [0, the first dimension size of x). Inputs: - **input_x** (Tensor) - The first input is a tensor whose data type is float16, float32 or int32. @@ -1076,7 +1076,7 @@ class Sub(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1115,7 +1115,7 @@ class Mul(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1154,7 +1154,7 @@ class SquaredDifference(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1341,7 +1341,7 @@ class Pow(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1453,11 +1453,11 @@ class HistogramFixedWidth(PrimitiveWithInfer): Args: dtype (string): An optional attribute. Must be one of the following types: "int32", "int64". Default: "int32". - nbins (int): Number of histogram bins, the type is positive integer. + nbins (int): The number of histogram bins, the type is a positive integer. Inputs: - **x** (Tensor) - Numeric Tensor. Must be one of the following types: int32, float32, float16. - - **range** (Tensor) - Must have the same type as x. Shape [2] Tensor of same dtype as x. + - **range** (Tensor) - Must has the same data type as `x`, and the shape is [2]. x <= range[0] will be mapped to hist[0], x >= range[1] will be mapped to hist[-1]. Outputs: @@ -1593,7 +1593,7 @@ class Erfc(PrimitiveWithInfer): Computes the complementary error function of `input_x` element-wise. Inputs: - - **input_x** (Tensor) - The input tensor. The data type mast be float16 or float32. + - **input_x** (Tensor) - The input tensor. The data type must be float16 or float32. Outputs: Tensor, has the same shape and dtype as the `input_x`. @@ -1627,7 +1627,7 @@ class Minimum(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1666,7 +1666,7 @@ class Maximum(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1705,7 +1705,7 @@ class RealDiv(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1744,13 +1744,13 @@ class Div(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or a bool or a tensor whose data type is number or bool. - **input_y** (Union[Tensor, Number, bool]) - When the first input is a tensor, The second input - could be a number or a bool, or a tensor whose data type is number or bool. When the first input + could be a number, a bool, or a tensor whose data type is number or bool. When the first input is a number or a bool, the second input should be a tensor whose data type is number or bool. Outputs: @@ -1758,7 +1758,7 @@ class Div(_MathBinaryOp): and the data type is the one with high precision or high digits among the two inputs. Raises: - ValueError: When `input_x` and `input_y` are not the same dtype. + ValueError: When `input_x` and `input_y` do not have the same dtype. Examples: >>> input_x = Tensor(np.array([-4.0, 5.0, 6.0]), mindspore.float32) @@ -1786,7 +1786,7 @@ class DivNoNan(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1799,7 +1799,7 @@ class DivNoNan(_MathBinaryOp): and the data type is the one with high precision or high digits among the two inputs. Raises: - ValueError: When `input_x` and `input_y` are not the same dtype. + ValueError: When `input_x` and `input_y` do not have the same dtype. Examples: >>> input_x = Tensor(np.array([-1.0, 0., 1.0, 5.0, 6.0]), mindspore.float32) @@ -1822,14 +1822,14 @@ class DivNoNan(_MathBinaryOp): class FloorDiv(_MathBinaryOp): """ - Divide the first input tensor by the second input tensor element-wise and rounds down to the closest integer. + Divide the first input tensor by the second input tensor element-wise and round down to the closest integer. Inputs of `input_x` and `input_y` comply with the implicit type conversion rules to make the data types consistent. The inputs must be two tensors or one tensor and one scalar. When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1860,7 +1860,7 @@ class TruncateDiv(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1890,7 +1890,7 @@ class TruncateMod(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -1918,7 +1918,7 @@ class Mod(_MathBinaryOp): Inputs of `input_x` and `input_y` comply with the implicit type conversion rules to make the data types consistent. The inputs must be two tensors or one tensor and one scalar. When the inputs are two tensors, both dtypes cannot be bool, and the shapes of them could be broadcast. When the inputs are one tensor - and one scalar, the scalar only could be a constant. + and one scalar, the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number]) - The first input is a number or a tensor whose data type is number. @@ -1953,7 +1953,7 @@ class Floor(PrimitiveWithInfer): Round a tensor down to the closest integer element-wise. Inputs: - - **input_x** (Tensor) - The input tensor. It's element data type must be float. + - **input_x** (Tensor) - The input tensor. Its element data type must be float. Outputs: Tensor, has the same shape as `input_x`. @@ -1979,14 +1979,14 @@ class Floor(PrimitiveWithInfer): class FloorMod(_MathBinaryOp): """ - Compute element-wise remainder of division. + Compute the remainder of division element-wise. Inputs of `input_x` and `input_y` comply with the implicit type conversion rules to make the data types consistent. The inputs must be two tensors or one tensor and one scalar. When the inputs are two tensors, dtypes of them cannot be both bool , and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2045,7 +2045,7 @@ class Xdivy(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2079,7 +2079,7 @@ class Xlogy(_MathBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2241,7 +2241,7 @@ class Equal(_LogicBinaryOp): Inputs of `input_x` and `input_y` comply with the implicit type conversion rules to make the data types consistent. The inputs must be two tensors or one tensor and one scalar. When the inputs are two tensors, the shapes of them could be broadcast. - When the inputs are one tensor and one scalar, the scalar only could be a constant. + When the inputs are one tensor and one scalar, the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number]) - The first input is a number or @@ -2356,7 +2356,7 @@ class NotEqual(_LogicBinaryOp): Inputs of `input_x` and `input_y` comply with the implicit type conversion rules to make the data types consistent. The inputs must be two tensors or one tensor and one scalar. When the inputs are two tensors, the shapes of them could be broadcast. - When the inputs are one tensor and one scalar, the scalar only could be a constant. + When the inputs are one tensor and one scalar, the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2393,7 +2393,7 @@ class Greater(_LogicBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2430,7 +2430,7 @@ class GreaterEqual(_LogicBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2467,7 +2467,7 @@ class Less(_LogicBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool, and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2504,7 +2504,7 @@ class LessEqual(_LogicBinaryOp): When the inputs are two tensors, dtypes of them cannot be both bool , and the shapes of them could be broadcast. When the inputs are one tensor and one scalar, - the scalar only could be a constant. + the scalar could only be a constant. Inputs: - **input_x** (Union[Tensor, Number, bool]) - The first input is a number or @@ -2570,7 +2570,7 @@ class LogicalAnd(_LogicBinaryOp): The inputs must be two tensors or one tensor and one bool. When the inputs are two tensors, the shapes of them could be broadcast, and the data types of them should be bool. - When the inputs are one tensor and one bool, the bool object only could be a constant, + When the inputs are one tensor and one bool, the bool object could only be a constant, and the data type of the tensor should be bool. Inputs: @@ -2601,7 +2601,7 @@ class LogicalOr(_LogicBinaryOp): The inputs must be two tensors or one tensor and one bool. When the inputs are two tensors, the shapes of them could be broadcast, and the data types of them should be bool. - When the inputs are one tensor and one bool, the bool object only could be a constant, + When the inputs are one tensor and one bool, the bool object could only be a constant, and the data type of the tensor should be bool. Inputs: @@ -2626,7 +2626,7 @@ class LogicalOr(_LogicBinaryOp): class IsNan(PrimitiveWithInfer): """ - Judging which elements are nan for each position + Judge which elements are nan for each position. Inputs: - **input_x** (Tensor) - The input tensor. @@ -2682,7 +2682,7 @@ class IsInf(PrimitiveWithInfer): class IsFinite(PrimitiveWithInfer): """ - Judging which elements are finite for each position + Judge which elements are finite for each position. Inputs: - **input_x** (Tensor) - The input tensor. @@ -2713,7 +2713,7 @@ class IsFinite(PrimitiveWithInfer): class FloatStatus(PrimitiveWithInfer): """ - Determine if the elements contains nan, inf or -inf. `0` for normal, `1` for overflow. + Determine if the elements contain Not a Number(NaN), infinite or negative infinite. 0 for normal, 1 for overflow. Inputs: - **input_x** (Tensor) - The input tensor. The data type must be float16 or float32. diff --git a/mindspore/ops/operations/nn_ops.py b/mindspore/ops/operations/nn_ops.py index 9d4c4bdcdd92dab4a040fa8dfff78e27ec0b251a..f72a8a9e767b0990ec36270d87dafab2e37e3e27 100644 --- a/mindspore/ops/operations/nn_ops.py +++ b/mindspore/ops/operations/nn_ops.py @@ -657,7 +657,7 @@ class FusedBatchNormEx(PrimitiveWithInfer): - **variance** (Tensor) - variance value, Tensor of shape :math:`(C,)`, data type: float32. Outputs: - Tuple of 6 Tensor, the normalized input, the updated parameters and reserve. + Tuple of 6 Tensors, the normalized input, the updated parameters and reserve. - **output_x** (Tensor) - The input of FusedBatchNormEx, same type and shape as the `input_x`. - **updated_scale** (Tensor) - Updated parameter scale, Tensor of shape :math:`(C,)`, data type: float32. @@ -870,13 +870,13 @@ class Conv2D(PrimitiveWithInfer): Args: out_channel (int): The dimension of the output. kernel_size (Union[int, tuple[int]]): The kernel size of the 2D convolution. - mode (int): 0 Math convolutiuon, 1 cross-correlation convolution , + mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution , 2 deconvolution, 3 depthwise convolution. Default: 1. - pad_mode (str): "valid", "same", "pad" the mode to fill padding. Default: "valid". - pad (Union(int, tuple[int])): The pad value to fill. Default: 0. If `pad` is one integer, the padding of - top, bottom, left and right is same, equal to pad. If `pad` is tuple with four integer, the padding - of top, bottom, left and right equal to pad[0], pad[1], pad[2], pad[3] with corresponding. - stride (Union(int, tuple[int])): The stride to apply conv filter. Default: 1. + pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid". + pad (Union(int, tuple[int])): The pad value to be filled. Default: 0. If `pad` is an integer, the paddings of + top, bottom, left and right are the same, equal to pad. If `pad` is a tuple of four integers, the + padding of top, bottom, left and right equal to pad[0], pad[1], pad[2], and pad[3] correspondingly. + stride (Union(int, tuple[int])): The stride to be applied to the convolution filter. Default: 1. dilation (Union(int, tuple[int])): Specify the space to use between kernel elements. Default: 1. group (int): Split input into groups. Default: 1. @@ -997,25 +997,26 @@ class DepthwiseConv2dNative(PrimitiveWithInfer): Given an input tensor of shape :math:`(N, C_{in}, H_{in}, W_{in})` where :math:`N` is the batch size and a filter tensor with kernel size :math:`(ks_{h}, ks_{w})`, containing :math:`C_{in} * \text{channel_multiplier}` convolutional filters of depth 1; it applies different filters to each input channel (channel_multiplier channels - for each with default value 1), then concatenates the results together. The output has + for each input channel has the default value 1), then concatenates the results together. The output has :math:`\text{in_channels} * \text{channel_multiplier}` channels. Args: - channel_multiplier (int): The multipiler for the original output conv. Its value must be greater than 0. - kernel_size (Union[int, tuple[int]]): The size of the conv kernel. - mode (int): 0 Math convolution, 1 cross-correlation convolution , + channel_multiplier (int): The multipiler for the original output convolution. Its value must be greater than 0. + kernel_size (Union[int, tuple[int]]): The size of the convolution kernel. + mode (int): Modes for different convolutions. 0 Math convolution, 1 cross-correlation convolution , 2 deconvolution, 3 depthwise convolution. Default: 3. - pad_mode (str): "valid", "same", "pad" the mode to fill padding. Default: "valid". - pad (Union[int, tuple[int]]): The pad value to fill. If `pad` is one integer, the padding of - top, bottom, left and right is same, equal to pad. If `pad` is tuple with four integer, the padding - of top, bottom, left and right equal to pad[0], pad[1], pad[2], pad[3] with corresponding. Default: 0. - stride (Union[int, tuple[int]]): The stride to apply conv filter. Default: 1. - dilation (Union[int, tuple[int]]): Specifies the dilation rate to use for dilated convolution. Default: 1. + pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid". + pad (Union[int, tuple[int]]): The pad value to be filled. If `pad` is an integer, the paddings of + top, bottom, left and right are the same, equal to pad. If `pad` is a tuple of four integers, the padding + of top, bottom, left and right equal to pad[0], pad[1], pad[2], and pad[3] correspondingly. Default: 0. + stride (Union[int, tuple[int]]): The stride to be applied to the convolution filter. Default: 1. + dilation (Union[int, tuple[int]]): Specifies the dilation rate to be used for the dilated convolution. + Default: 1. group (int): Splits input into groups. Default: 1. Inputs: - **input** (Tensor) - Tensor of shape :math:`(N, C_{in}, H_{in}, W_{in})`. - - **weight** (Tensor) - Set size of kernel is :math:`(K_1, K_2)`, then the shape is + - **weight** (Tensor) - Set the size of kernel as :math:`(K_1, K_2)`, then the shape is :math:`(K, C_{in}, K_1, K_2)`, `K` must be 1. Outputs: @@ -1398,14 +1399,15 @@ class Conv2DBackpropInput(PrimitiveWithInfer): Args: out_channel (int): The dimensionality of the output space. kernel_size (Union[int, tuple[int]]): The size of the convolution window. - pad_mode (str): "valid", "same", "pad" the mode to fill padding. Default: "valid". - pad (Union[int, tuple[int]]): The pad value to fill. Default: 0. If `pad` is one integer, the padding of - top, bottom, left and right is same, equal to pad. If `pad` is tuple with four integer, the padding - of top, bottom, left and right equal to pad[0], pad[1], pad[2], pad[3] with corresponding. - mode (int): 0 Math convolutiuon, 1 cross-correlation convolution , + pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid". + pad (Union[int, tuple[int]]): The pad value to be filled. Default: 0. If `pad` is an integer, the paddings of + top, bottom, left and right are the same, equal to pad. If `pad` is a tuple of four integers, the + padding of top, bottom, left and right equal to pad[0], pad[1], pad[2], and pad[3] correspondingly. + mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution , 2 deconvolution, 3 depthwise convolution. Default: 1. - stride (Union[int. tuple[int]]): The stride to apply conv filter. Default: 1. - dilation (Union[int. tuple[int]]): Specifies the dilation rate to use for dilated convolution. Default: 1. + stride (Union[int. tuple[int]]): The stride to be applied to the convolution filter. Default: 1. + dilation (Union[int. tuple[int]]): Specifies the dilation rate to be used for the dilated convolution. + Default: 1. group (int): Splits input into groups. Default: 1. Returns: @@ -1842,7 +1844,7 @@ class L2Loss(PrimitiveWithInfer): class DataFormatDimMap(PrimitiveWithInfer): """ - Returns the dimension index in the destination data format given the one in the source data format. + Returns the dimension index in the destination data format given in the source data format. Args: src_format (string): An optional value for source data format. Default: 'NHWC'. @@ -2336,7 +2338,7 @@ class DropoutDoMask(PrimitiveWithInfer): Inputs: - **input_x** (Tensor) - The input tensor. - **mask** (Tensor) - The mask to be applied on `input_x`, which is the output of `DropoutGenMask`. And the - shape of `input_x` must be same as the value of `DropoutGenMask`'s input `shape`. If input wrong `mask`, + shape of `input_x` must be the same as the value of `DropoutGenMask`'s input `shape`. If input wrong `mask`, the output of `DropoutDoMask` are unpredictable. - **keep_prob** (Tensor) - The keep rate, between 0 and 1, e.g. keep_prob = 0.9, means dropping out 10% of input units. The value of `keep_prob` is the same as the input `keep_prob` of @@ -2494,10 +2496,10 @@ class Gelu(PrimitiveWithInfer): Gaussian Error Linear Units activation function. GeLU is described in the paper `Gaussian Error Linear Units (GELUs) `_. - And also please refer to `BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. + And also please refer to `BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding `_. - Defined as follows: + Gelu is defined as follows: .. math:: \text{output} = 0.5 * x * (1 + erf(x / \sqrt{2})), @@ -2505,7 +2507,7 @@ class Gelu(PrimitiveWithInfer): where :math:`erf` is the "Gauss error function" . Inputs: - - **input_x** (Tensor) - Input to compute the Gelu. With data type of float16 or float32. + - **input_x** (Tensor) - Input to compute the Gelu with data type of float16 or float32. Outputs: Tensor, with the same type and shape as input. @@ -2534,8 +2536,8 @@ class GetNext(PrimitiveWithInfer): Returns the next element in the dataset queue. Note: - GetNext op needs to be associated with network and also depends on the init_dataset interface, - it can't be used directly as a single op. + The GetNext operation needs to be associated with network and it also depends on the init_dataset interface, + it can't be used directly as a single operation. For details, please refer to `nn.DataWrapper` source code. Args: @@ -3057,7 +3059,7 @@ class Adam(PrimitiveWithInfer): class FusedSparseAdam(PrimitiveWithInfer): r""" - Merge the duplicate value of the gradient and then updates parameters by Adaptive Moment Estimation (Adam) + Merge the duplicate value of the gradient and then update parameters by Adaptive Moment Estimation (Adam) algorithm. This operator is used when the gradient is sparse. The Adam algorithm is proposed in `Adam: A Method for Stochastic Optimization `_. @@ -3092,22 +3094,22 @@ class FusedSparseAdam(PrimitiveWithInfer): If true, update the gradients without using NAG. Default: False. Inputs: - - **var** (Parameter) - Parameters to be updated. With float32 data type. - - **m** (Parameter) - The 1st moment vector in the updating formula, has the same type as `var`. With - float32 data type. - - **v** (Parameter) - The 2nd moment vector in the updating formula. Mean square gradients - with the same type as `var`. With float32 data type. - - **beta1_power** (Tensor) - :math:`beta_1^t` in the updating formula. With float32 data type. - - **beta2_power** (Tensor) - :math:`beta_2^t` in the updating formula. With float32 data type. + - **var** (Parameter) - Parameters to be updated with float32 data type. + - **m** (Parameter) - The 1st moment vector in the updating formula, has the same type as `var` with + float32 data type. + - **v** (Parameter) - The 2nd moment vector in the updating formula. Mean square gradients, has the same type as + `var` with float32 data type. + - **beta1_power** (Tensor) - :math:`beta_1^t` in the updating formula with float32 data type. + - **beta2_power** (Tensor) - :math:`beta_2^t` in the updating formula with float32 data type. - **lr** (Tensor) - :math:`l` in the updating formula. With float32 data type. - - **beta1** (Tensor) - The exponential decay rate for the 1st moment estimations. With float32 data type. - - **beta2** (Tensor) - The exponential decay rate for the 2nd moment estimations. With float32 data type. - - **epsilon** (Tensor) - Term added to the denominator to improve numerical stability. With float32 data type. - - **gradient** (Tensor) - Gradient value. With float32 data type. - - **indices** (Tensor) - Gradient indices. With int32 data type. + - **beta1** (Tensor) - The exponential decay rate for the 1st moment estimations with float32 data type. + - **beta2** (Tensor) - The exponential decay rate for the 2nd moment estimations with float32 data type. + - **epsilon** (Tensor) - Term added to the denominator to improve numerical stability with float32 data type. + - **gradient** (Tensor) - Gradient value with float32 data type. + - **indices** (Tensor) - Gradient indices with int32 data type. Outputs: - Tuple of 3 Tensor, this operator will update the input parameters directly, the outputs are useless. + Tuple of 3 Tensors, this operator will update the input parameters directly, the outputs are useless. - **var** (Tensor) - A Tensor with shape (1,). - **m** (Tensor) - A Tensor with shape (1,). @@ -3189,7 +3191,7 @@ class FusedSparseAdam(PrimitiveWithInfer): class FusedSparseLazyAdam(PrimitiveWithInfer): r""" - Merge the duplicate value of the gradient and then updates parameters by Adaptive Moment Estimation (Adam) + Merge the duplicate value of the gradient and then update parameters by Adaptive Moment Estimation (Adam) algorithm. This operator is used when the gradient is sparse. The behavior is not equivalent to the original Adam algorithm, as only the current indices parameters will be updated. @@ -3225,22 +3227,22 @@ class FusedSparseLazyAdam(PrimitiveWithInfer): If true, update the gradients without using NAG. Default: False. Inputs: - - **var** (Parameter) - Parameters to be updated. With float32 data type. - - **m** (Parameter) - The 1st moment vector in the updating formula, has the same type as `var`. With - float32 data type. - - **v** (Parameter) - The 2nd moment vector in the updating formula. Mean square gradients - with the same type as `var`. With float32 data type. - - **beta1_power** (Tensor) - :math:`beta_1^t` in the updating formula. With float32 data type. - - **beta2_power** (Tensor) - :math:`beta_2^t` in the updating formula. With float32 data type. - - **lr** (Tensor) - :math:`l` in the updating formula. With float32 data type. - - **beta1** (Tensor) - The exponential decay rate for the 1st moment estimations. With float32 data type. - - **beta2** (Tensor) - The exponential decay rate for the 2nd moment estimations. With float32 data type. - - **epsilon** (Tensor) - Term added to the denominator to improve numerical stability. With float32 data type. - - **gradient** (Tensor) - Gradient value. With float32 data type. - - **indices** (Tensor) - Gradient indices. With int32 data type. + - **var** (Parameter) - Parameters to be updated with float32 data type. + - **m** (Parameter) - The 1st moment vector in the updating formula, has the same type as `var` with + float32 data type. + - **v** (Parameter) - The 2nd moment vector in the updating formula. Mean square gradients, has the same type as + `var` with float32 data type. + - **beta1_power** (Tensor) - :math:`beta_1^t` in the updating formula with float32 data type. + - **beta2_power** (Tensor) - :math:`beta_2^t` in the updating formula with float32 data type. + - **lr** (Tensor) - :math:`l` in the updating formula with float32 data type. + - **beta1** (Tensor) - The exponential decay rate for the 1st moment estimations with float32 data type. + - **beta2** (Tensor) - The exponential decay rate for the 2nd moment estimations with float32 data type. + - **epsilon** (Tensor) - Term added to the denominator to improve numerical stability with float32 data type. + - **gradient** (Tensor) - Gradient value with float32 data type. + - **indices** (Tensor) - Gradient indices with int32 data type. Outputs: - Tuple of 3 Tensor, this operator will update the input parameters directly, the outputs are useless. + Tuple of 3 Tensors, this operator will update the input parameters directly, the outputs are useless. - **var** (Tensor) - A Tensor with shape (1,). - **m** (Tensor) - A Tensor with shape (1,). @@ -3418,7 +3420,7 @@ class FusedSparseFtrl(PrimitiveWithInfer): class FusedSparseProximalAdagrad(PrimitiveWithInfer): r""" - Merge the duplicate value of the gradient and then Updates relevant entries according to the proximal adagrad + Merge the duplicate value of the gradient and then update relevant entries according to the proximal adagrad algorithm. .. math:: @@ -3434,7 +3436,7 @@ class FusedSparseProximalAdagrad(PrimitiveWithInfer): RuntimeError exception will be thrown when the data type conversion of Parameter is required. Args: - use_locking (bool): If true, the var and accumulation tensors will be protected from being updated. + use_locking (bool): If true, the variable and accumulation tensors will be protected from being updated. Default: False. Inputs: @@ -3448,7 +3450,7 @@ class FusedSparseProximalAdagrad(PrimitiveWithInfer): must be int32. Outputs: - Tuple of 2 Tensor, this operator will update the input parameters directly, the outputs are useless. + Tuple of 2 Tensors, this operator will update the input parameters directly, the outputs are useless. - **var** (Tensor) - A Tensor with shape (1,). - **accum** (Tensor) - A Tensor with shape (1,). @@ -3524,9 +3526,9 @@ class KLDivLoss(PrimitiveWithInfer): .. math:: \ell(x, y) = \begin{cases} - L, & \text{if reduction} = \text{'none';}\\ - \operatorname{mean}(L), & \text{if reduction} = \text{'mean';}\\ - \operatorname{sum}(L), & \text{if reduction} = \text{'sum'.} + L, & \text{if reduction} = \text{`none';}\\ + \operatorname{mean}(L), & \text{if reduction} = \text{`mean';}\\ + \operatorname{sum}(L), & \text{if reduction} = \text{`sum'.} \end{cases} Args: @@ -3535,10 +3537,10 @@ class KLDivLoss(PrimitiveWithInfer): Inputs: - **input_x** (Tensor) - The input Tensor. The data type must be float32. - - **input_y** (Tensor) - The label Tensor which has same shape as `input_x`. The data type must be float32. + - **input_y** (Tensor) - The label Tensor which has the same shape as `input_x`. The data type must be float32. Outputs: - Tensor or Scalar, if `reduction` is 'none', then output is a tensor and same shape as `input_x`. + Tensor or Scalar, if `reduction` is 'none', then output is a tensor and has the same shape as `input_x`. Otherwise it is a scalar. Examples: @@ -5151,15 +5153,15 @@ class SparseApplyFtrlV2(PrimitiveWithInfer): class ConfusionMulGrad(PrimitiveWithInfer): """ - `output0` is the result of which input0 dot multily input1. + `output0` is the dot product result of input0 and input1. - `output1` is the result of which input0 dot multily input1, then reducesum it. + `output1` is the dot product result of input0 and input1, then apply the reducesum operation on it. Args: axis (Union[int, tuple[int], list[int]]): The dimensions to reduce. Default:(), reduce all dimensions. Only constant value is allowed. keep_dims (bool): - - If true, keep these reduced dimensions and the length is 1. + - If true, keep these reduced dimensions and the length as 1. - If false, don't keep these dimensions. Default:False. Inputs: @@ -5167,8 +5169,8 @@ class ConfusionMulGrad(PrimitiveWithInfer): - **input_1** (Tensor) - The input Tensor. - **input_2** (Tensor) - The input Tensor. - outputs: - - **output_0** (Tensor) - The same shape with `input0`. + Outputs: + - **output_0** (Tensor) - The same shape as `input0`. - **output_1** (Tensor) - If axis is (), and keep_dims is false, the output is a 0-D array representing @@ -5462,7 +5464,7 @@ class BasicLSTMCell(PrimitiveWithInfer): - **w** (Tensor) - Weight. Tensor of shape (`input_size + hidden_size`, `4 x hidden_size`). The data type must be float16 or float32. - **b** (Tensor) - Bias. Tensor of shape (`4 x hidden_size`). - The data type must be same as `c`. + The data type must be the same as `c`. Outputs: - **ct** (Tensor) - Forward :math:`c_t` cache at moment `t`. Tensor of shape (`batch_size`, `hidden_size`). @@ -5532,18 +5534,18 @@ class BasicLSTMCell(PrimitiveWithInfer): class InTopK(PrimitiveWithInfer): r""" - Says whether the targets are in the top `k` predictions. + Whether the targets are in the top `k` predictions. Args: - k (int): Special the number of top elements to look at for computing precision. + k (int): Specify the number of top elements to be used for computing precision. Inputs: - - **x1** (Tensor) - A 2D Tensor define the predictions of a batch of samples with float16 or float32 data type. - - **x2** (Tensor) - A 1D Tensor define the labels of a batch of samples with int32 data type. + - **x1** (Tensor) - A 2D Tensor defines the predictions of a batch of samples with float16 or float32 data type. + - **x2** (Tensor) - A 1D Tensor defines the labels of a batch of samples with int32 data type. Outputs: - Tensor, which is 1 dimension of type bool and has same shape with `x2`. for label of sample `i` in `x2`, - if label in first `k` predictions for sample `i` in `x1`, then the value is True else False. + Tensor has 1 dimension of type bool and the same shape with `x2`. For labeling sample `i` in `x2`, + if the label in the first `k` predictions for sample `i` is in `x1`, then the value is True, otherwise False. Examples: >>> x1 = Tensor(np.array([[1, 8, 5, 2, 7], [4, 9, 1, 3, 5]]), mindspore.float32) diff --git a/mindspore/ops/operations/other_ops.py b/mindspore/ops/operations/other_ops.py index 4ac5b8b058a89939dfe9943a5b77a47be5897efa..57503ab38b420e85f5ae56c83497d221c05eafa2 100644 --- a/mindspore/ops/operations/other_ops.py +++ b/mindspore/ops/operations/other_ops.py @@ -244,7 +244,7 @@ class IOU(PrimitiveWithInfer): Args: mode (string): The mode is used to specify the calculation method, - now support 'iou' (intersection over union) or 'iof' + now supporting 'iou' (intersection over union) or 'iof' (intersection over foreground) mode. Default: 'iou'. Inputs: @@ -350,7 +350,7 @@ class Partial(Primitive): class Depend(Primitive): """ - Depend is used for process side-effect operations. + Depend is used for processing side-effect operations. Inputs: - **value** (Tensor) - the real value to return for depend operator. diff --git a/mindspore/ops/operations/random_ops.py b/mindspore/ops/operations/random_ops.py index 49566089621cc42a4ed8d3da09757ee28323b0f8..7f137a48166980b2691f062f257d92487f170754 100644 --- a/mindspore/ops/operations/random_ops.py +++ b/mindspore/ops/operations/random_ops.py @@ -131,9 +131,9 @@ class Gamma(PrimitiveWithInfer): Inputs: - **shape** (tuple) - The shape of random tensor to be generated. Only constant value is allowed. - **alpha** (Tensor) - The α distribution parameter. - It is also known as the shape parameter. With float32 data type. + It is also known as the shape parameter with float32 data type. - **beta** (Tensor) - The β distribution parameter. - It is also known as the scale parameter. With float32 data type. + It is also known as the scale parameter with float32 data type. Outputs: Tensor. The shape should be the broadcasted shape of Input "shape" and shapes of alpha and beta. diff --git a/mindspore/parallel/algo_parameter_config.py b/mindspore/parallel/algo_parameter_config.py index 5c13c13153a558b1037d228386979d8c9059f219..462ee0b371a88edaf5f8e48262ca107a1924759e 100644 --- a/mindspore/parallel/algo_parameter_config.py +++ b/mindspore/parallel/algo_parameter_config.py @@ -130,14 +130,14 @@ def set_algo_parameters(**kwargs): Set algo parameter config. Note: - Attribute name is needed. + The attribute name is required. Args: - tensor_slice_align_enable (bool): Whether checking tensor slice shape for MatMul. Default: False + tensor_slice_align_enable (bool): Whether to check the shape of tensor slice of MatMul. Default: False tensor_slice_align_size (int): The minimum tensor slice shape of MatMul, the value must be in [1, 1024]. Default: 16 fully_use_devices (bool): Whether ONLY generating strategies that fully use all available devices. Default: True - elementwise_op_strategy_follow (bool): Whether the elementwise operator have the same strategies as its + elementwise_op_strategy_follow (bool): Whether the elementwise operator has the same strategies as its subsequent operators. Default: False Raises: @@ -155,7 +155,7 @@ def get_algo_parameters(attr_key): Get algo parameter config attributes. Note: - Return value according to the attribute value. + Returns the specified attribute value. Args: attr_key (str): The key of the attribute.