diff --git a/mindspore/nn/layer/normalization.py b/mindspore/nn/layer/normalization.py index b286eaae1b00909a92f956393fcba2a39d5b1496..4aafaf031e1ba271f8092d390ab7c01078c8e5a5 100644 --- a/mindspore/nn/layer/normalization.py +++ b/mindspore/nn/layer/normalization.py @@ -20,7 +20,7 @@ from mindspore.common.initializer import initializer from mindspore.common.tensor import Tensor import mindspore.common.dtype as mstype import mindspore.context as context -from mindspore._checkparam import check_int_positive, check_bool, check_typename +from mindspore._checkparam import check_int_positive, check_bool, check_typename from mindspore._extends import cell_attr_register from ..cell import Cell @@ -293,7 +293,7 @@ class LayerNorm(Cell): class GroupNorm(Cell): r""" - Group Normalization over a mini-batch of inputs. + Group Normalization over a mini-batch of inputs. Group normalization is widely used in recurrent neural networks. It applies normalization over a mini-batch of inputs for each single training case as described