未验证 提交 7bfd799d 编写于 作者: C cnn 提交者: GitHub

Release 2.0rc cherry pick api rename #28108 (#28184)

* rename count_include_pad-->exclusive  return_indices-->return_mask

* remove track_running_stats

* fix typo.

* rename xxxd-->xxxxD

* solve conflicts
上级 b04c55ef
...@@ -63,10 +63,7 @@ class TestLayer(fluid.dygraph.Layer): ...@@ -63,10 +63,7 @@ class TestLayer(fluid.dygraph.Layer):
bias_attr=False) bias_attr=False)
self._sync_batch_norm2 = SyncBatchNorm( self._sync_batch_norm2 = SyncBatchNorm(
num_filters, num_filters, weight_attr=False, bias_attr=False)
weight_attr=False,
bias_attr=False,
track_running_stats=False)
def forward(self, inputs): def forward(self, inputs):
y = self._conv(inputs) y = self._conv(inputs)
......
...@@ -150,7 +150,7 @@ class TestAdaptiveMaxPool2DAPI(unittest.TestCase): ...@@ -150,7 +150,7 @@ class TestAdaptiveMaxPool2DAPI(unittest.TestCase):
x = paddle.to_tensor(self.x_np) x = paddle.to_tensor(self.x_np)
out_1 = paddle.nn.functional.adaptive_max_pool2d( out_1 = paddle.nn.functional.adaptive_max_pool2d(
x=x, return_indices=False, output_size=[3, 3]) x=x, return_mask=False, output_size=[3, 3])
out_2 = paddle.nn.functional.adaptive_max_pool2d(x=x, output_size=5) out_2 = paddle.nn.functional.adaptive_max_pool2d(x=x, output_size=5)
......
...@@ -148,11 +148,7 @@ class TestPool1D_API(unittest.TestCase): ...@@ -148,11 +148,7 @@ class TestPool1D_API(unittest.TestCase):
input_np = np.random.random([2, 3, 32]).astype("float32") input_np = np.random.random([2, 3, 32]).astype("float32")
input = fluid.dygraph.to_variable(input_np) input = fluid.dygraph.to_variable(input_np)
result = F.avg_pool1d( result = F.avg_pool1d(
input, input, kernel_size=2, stride=2, padding=[1], exclusive=True)
kernel_size=2,
stride=2,
padding=[1],
count_include_pad=True)
result_np = avg_pool1D_forward_naive( result_np = avg_pool1D_forward_naive(
input_np, ksize=[2], strides=[2], paddings=[1], exclusive=False) input_np, ksize=[2], strides=[2], paddings=[1], exclusive=False)
...@@ -160,7 +156,8 @@ class TestPool1D_API(unittest.TestCase): ...@@ -160,7 +156,8 @@ class TestPool1D_API(unittest.TestCase):
self.assertTrue(np.allclose(result.numpy(), result_np)) self.assertTrue(np.allclose(result.numpy(), result_np))
avg_pool1d_dg = paddle.nn.AvgPool1D( avg_pool1d_dg = paddle.nn.AvgPool1D(
kernel_size=2, stride=None, padding=1, count_include_pad=True) kernel_size=2, stride=None, padding=1, exclusive=True)
result = avg_pool1d_dg(input) result = avg_pool1d_dg(input)
self.assertTrue(np.allclose(result.numpy(), result_np)) self.assertTrue(np.allclose(result.numpy(), result_np))
...@@ -200,7 +197,7 @@ class TestPool1D_API(unittest.TestCase): ...@@ -200,7 +197,7 @@ class TestPool1D_API(unittest.TestCase):
input_np = np.random.random([2, 3, 32]).astype("float32") input_np = np.random.random([2, 3, 32]).astype("float32")
input = fluid.dygraph.to_variable(input_np) input = fluid.dygraph.to_variable(input_np)
result, index = F.max_pool1d( result, index = F.max_pool1d(
input, kernel_size=2, stride=2, padding=0, return_indices=True) input, kernel_size=2, stride=2, padding=0, return_mask=True)
result_np = max_pool1D_forward_naive( result_np = max_pool1D_forward_naive(
input_np, ksize=[2], strides=[2], paddings=[0]) input_np, ksize=[2], strides=[2], paddings=[0])
......
...@@ -134,7 +134,7 @@ class TestPool2D_API(unittest.TestCase): ...@@ -134,7 +134,7 @@ class TestPool2D_API(unittest.TestCase):
input_np = np.random.random([2, 3, 32, 32]).astype("float32") input_np = np.random.random([2, 3, 32, 32]).astype("float32")
input = fluid.dygraph.to_variable(input_np) input = fluid.dygraph.to_variable(input_np)
result = max_pool2d( result = max_pool2d(
input, kernel_size=2, stride=2, padding=0, return_indices=False) input, kernel_size=2, stride=2, padding=0, return_mask=False)
result_np = pool2D_forward_naive( result_np = pool2D_forward_naive(
input_np, input_np,
...@@ -159,7 +159,7 @@ class TestPool2D_API(unittest.TestCase): ...@@ -159,7 +159,7 @@ class TestPool2D_API(unittest.TestCase):
kernel_size=2, kernel_size=2,
stride=2, stride=2,
padding=0, padding=0,
return_indices=False, return_mask=False,
data_format="NHWC") data_format="NHWC")
result_np = pool2D_forward_naive( result_np = pool2D_forward_naive(
...@@ -222,7 +222,7 @@ class TestPool2D_API(unittest.TestCase): ...@@ -222,7 +222,7 @@ class TestPool2D_API(unittest.TestCase):
kernel_size=2, kernel_size=2,
stride=None, stride=None,
padding="SAME", padding="SAME",
return_indices=True) return_mask=True)
result_np = pool2D_forward_naive( result_np = pool2D_forward_naive(
input_np, input_np,
...@@ -269,7 +269,7 @@ class TestPool2D_API(unittest.TestCase): ...@@ -269,7 +269,7 @@ class TestPool2D_API(unittest.TestCase):
kernel_size=2, kernel_size=2,
stride=2, stride=2,
padding=padding, padding=padding,
return_indices=False) return_mask=False)
result_np = pool2D_forward_naive( result_np = pool2D_forward_naive(
input_np, input_np,
...@@ -490,7 +490,7 @@ class TestPool2DError_API(unittest.TestCase): ...@@ -490,7 +490,7 @@ class TestPool2DError_API(unittest.TestCase):
padding=0, padding=0,
ceil_mode=False, ceil_mode=False,
data_format='NHWC', data_format='NHWC',
return_indices=True) return_mask=True)
self.assertRaises(ValueError, run9) self.assertRaises(ValueError, run9)
......
...@@ -83,7 +83,7 @@ class TestPool3D_API(unittest.TestCase): ...@@ -83,7 +83,7 @@ class TestPool3D_API(unittest.TestCase):
stride=2, stride=2,
padding=1, padding=1,
ceil_mode=False, ceil_mode=False,
count_include_pad=True) exclusive=True)
result_np = avg_pool3D_forward_naive( result_np = avg_pool3D_forward_naive(
input_np, input_np,
...@@ -100,7 +100,7 @@ class TestPool3D_API(unittest.TestCase): ...@@ -100,7 +100,7 @@ class TestPool3D_API(unittest.TestCase):
stride=None, stride=None,
padding=1, padding=1,
ceil_mode=False, ceil_mode=False,
count_include_pad=True) exclusive=True)
result = avg_pool3d_dg(input) result = avg_pool3d_dg(input)
self.assertTrue(np.allclose(result.numpy(), result_np)) self.assertTrue(np.allclose(result.numpy(), result_np))
...@@ -175,7 +175,7 @@ class TestPool3D_API(unittest.TestCase): ...@@ -175,7 +175,7 @@ class TestPool3D_API(unittest.TestCase):
stride=2, stride=2,
padding=0, padding=0,
data_format="NDHWC", data_format="NDHWC",
return_indices=False) return_mask=False)
result_np = pool3D_forward_naive( result_np = pool3D_forward_naive(
input_np, input_np,
...@@ -239,7 +239,7 @@ class TestPool3D_API(unittest.TestCase): ...@@ -239,7 +239,7 @@ class TestPool3D_API(unittest.TestCase):
kernel_size=2, kernel_size=2,
stride=None, stride=None,
padding="SAME", padding="SAME",
return_indices=True) return_mask=True)
result_np = pool3D_forward_naive( result_np = pool3D_forward_naive(
input_np, input_np,
...@@ -467,7 +467,7 @@ class TestPool3DError_API(unittest.TestCase): ...@@ -467,7 +467,7 @@ class TestPool3DError_API(unittest.TestCase):
stride=2, stride=2,
padding=0, padding=0,
data_format='NDHWC', data_format='NDHWC',
return_indices=True) return_mask=True)
self.assertRaises(ValueError, run10) self.assertRaises(ValueError, run10)
......
...@@ -73,7 +73,6 @@ class _InstanceNormBase(layers.Layer): ...@@ -73,7 +73,6 @@ class _InstanceNormBase(layers.Layer):
momentum=0.9, momentum=0.9,
weight_attr=None, weight_attr=None,
bias_attr=None, bias_attr=None,
track_running_stats=False,
data_format="NCHW", data_format="NCHW",
name=None): name=None):
super(_InstanceNormBase, self).__init__() super(_InstanceNormBase, self).__init__()
...@@ -135,9 +134,6 @@ class InstanceNorm1D(_InstanceNormBase): ...@@ -135,9 +134,6 @@ class InstanceNorm1D(_InstanceNormBase):
epsilon(float, optional): A value added to the denominator for epsilon(float, optional): A value added to the denominator for
numerical stability. Default is 1e-5. numerical stability. Default is 1e-5.
momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9. momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9.
track_running_stats(bool, optional): Whether to use global mean and
variance. In train mode, when setting track_running_stats True, the global mean
and variance are also used during train period. Default: False.
weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale` weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale`
of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm
will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr. will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr.
...@@ -159,9 +155,6 @@ class InstanceNorm1D(_InstanceNormBase): ...@@ -159,9 +155,6 @@ class InstanceNorm1D(_InstanceNormBase):
Returns: Returns:
None. None.
**Note**:
Momentum and track_running_stats is not effective. The next version will fix the problem .
Examples: Examples:
...@@ -214,9 +207,6 @@ class InstanceNorm2D(_InstanceNormBase): ...@@ -214,9 +207,6 @@ class InstanceNorm2D(_InstanceNormBase):
epsilon(float, optional): A value added to the denominator for epsilon(float, optional): A value added to the denominator for
numerical stability. Default is 1e-5. numerical stability. Default is 1e-5.
momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9. momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9.
track_running_stats(bool, optional): Whether to use global mean and
variance. In train mode, when setting track_running_stats True, the global mean
and variance are also used during train period. Default: False.
weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale` weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale`
of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm
will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr. will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr.
...@@ -237,8 +227,6 @@ class InstanceNorm2D(_InstanceNormBase): ...@@ -237,8 +227,6 @@ class InstanceNorm2D(_InstanceNormBase):
Returns: Returns:
None. None.
**Note**:
Momentum and track_running_stats is not effective. The next version will fix the problem .
Examples: Examples:
...@@ -290,9 +278,6 @@ class InstanceNorm3D(_InstanceNormBase): ...@@ -290,9 +278,6 @@ class InstanceNorm3D(_InstanceNormBase):
epsilon(float, optional): A value added to the denominator for epsilon(float, optional): A value added to the denominator for
numerical stability. Default is 1e-5. numerical stability. Default is 1e-5.
momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9. momentum(float, optional): The value used for the moving_mean and moving_var computation. Default: 0.9.
track_running_stats(bool, optional): Whether to use global mean and
variance. In train mode, when setting track_running_stats True, the global mean
and variance are also used during train period. Default: False.
weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale` weight_attr(ParamAttr|bool, optional): The parameter attribute for Parameter `scale`
of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm of instance_norm. If it is set to None or one attribute of ParamAttr, instance_norm
will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr. will create ParamAttr as weight_attr, the name of scale can be set in ParamAttr.
...@@ -313,8 +298,6 @@ class InstanceNorm3D(_InstanceNormBase): ...@@ -313,8 +298,6 @@ class InstanceNorm3D(_InstanceNormBase):
Returns: Returns:
None. None.
**Note**:
Momentum and track_running_stats is not effective. The next version will fix the problem .
Examples: Examples:
...@@ -570,7 +553,6 @@ class _BatchNormBase(layers.Layer): ...@@ -570,7 +553,6 @@ class _BatchNormBase(layers.Layer):
weight_attr=None, weight_attr=None,
bias_attr=None, bias_attr=None,
data_format='NCHW', data_format='NCHW',
track_running_stats=True,
name=None): name=None):
super(_BatchNormBase, self).__init__() super(_BatchNormBase, self).__init__()
self._num_features = num_features self._num_features = num_features
...@@ -636,7 +618,6 @@ class _BatchNormBase(layers.Layer): ...@@ -636,7 +618,6 @@ class _BatchNormBase(layers.Layer):
self._momentum = momentum self._momentum = momentum
self._epsilon = epsilon self._epsilon = epsilon
self._fuse_with_relu = False self._fuse_with_relu = False
self._track_running_stats = track_running_stats
self._name = name self._name = name
def _check_input_dim(self, input): def _check_input_dim(self, input):
...@@ -651,11 +632,7 @@ class _BatchNormBase(layers.Layer): ...@@ -651,11 +632,7 @@ class _BatchNormBase(layers.Layer):
self._check_input_dim(input) self._check_input_dim(input)
if not self.training and not self._track_running_stats: if self.training:
raise ValueError(
'When inference, expected track_running_stats is True.')
if self.training and not self._track_running_stats:
warnings.warn( warnings.warn(
"When training, we now always track global mean and variance.") "When training, we now always track global mean and variance.")
...@@ -720,9 +697,6 @@ class BatchNorm1D(_BatchNormBase): ...@@ -720,9 +697,6 @@ class BatchNorm1D(_BatchNormBase):
will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable. will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable.
If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None. If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None.
data_format(str, optional): Specify the input data format, may be "NC", "NCL" or "NLC". Defalut "NCL". data_format(str, optional): Specify the input data format, may be "NC", "NCL" or "NLC". Defalut "NCL".
track_running_stats(bool, optional): Whether to use global mean and variance. In train period,
True will track global mean and variance used for inference. When inference, track_running_stats must be
True. Default: True.
name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`.. name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`..
Shape: Shape:
...@@ -732,9 +706,6 @@ class BatchNorm1D(_BatchNormBase): ...@@ -732,9 +706,6 @@ class BatchNorm1D(_BatchNormBase):
Returns: Returns:
None. None.
**Note**:
Now track_running_stats is actucal always true. The next version will fix the problem .
Examples: Examples:
...@@ -817,9 +788,6 @@ class BatchNorm2D(_BatchNormBase): ...@@ -817,9 +788,6 @@ class BatchNorm2D(_BatchNormBase):
will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable. will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable.
If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None. If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None.
data_format(str, optional): Specify the input data format, the data format can be "NCHW" or "NHWC". Default: NCHW. data_format(str, optional): Specify the input data format, the data format can be "NCHW" or "NHWC". Default: NCHW.
track_running_stats(bool, optional): Whether to use global mean and variance. In train period,
True will track global mean and variance used for inference. When inference, track_running_stats must be
True. Default: True.
name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`.. name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`..
Shape: Shape:
...@@ -830,9 +798,6 @@ class BatchNorm2D(_BatchNormBase): ...@@ -830,9 +798,6 @@ class BatchNorm2D(_BatchNormBase):
Returns: Returns:
None None
**Note**:
Now track_running_stats is actucal always true. The next version will fix the problem .
Examples: Examples:
.. code-block:: python .. code-block:: python
...@@ -912,9 +877,6 @@ class BatchNorm3D(_BatchNormBase): ...@@ -912,9 +877,6 @@ class BatchNorm3D(_BatchNormBase):
will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable. will create ParamAttr as bias_attr. If it is set to Fasle, the weight is not learnable.
If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None. If the Initializer of the bias_attr is not set, the bias is initialized zero. Default: None.
data_format(str, optional): Specify the input data format, the data format can be "NCDHW" or "NDHWC. Default: NCDHW. data_format(str, optional): Specify the input data format, the data format can be "NCDHW" or "NDHWC. Default: NCDHW.
track_running_stats(bool, optional): Whether to use global mean and variance. In train period,
True will track global mean and variance used for inference. When inference, track_running_stats must be
True. Default: True.
name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`.. name(str, optional): Name for the BatchNorm, default is None. For more information, please refer to :ref:`api_guide_Name`..
Shape: Shape:
...@@ -925,9 +887,6 @@ class BatchNorm3D(_BatchNormBase): ...@@ -925,9 +887,6 @@ class BatchNorm3D(_BatchNormBase):
Returns: Returns:
None None
**Note**:
Now track_running_stats is actucal always true. The next version will fix the problem .
Examples: Examples:
.. code-block:: python .. code-block:: python
...@@ -1024,8 +983,6 @@ class SyncBatchNorm(_BatchNormBase): ...@@ -1024,8 +983,6 @@ class SyncBatchNorm(_BatchNormBase):
will create ParamAttr as bias_attr. If the Initializer of the bias_attr will create ParamAttr as bias_attr. If the Initializer of the bias_attr
is not set, the bias is initialized zero. If it is set to False, this layer will not is not set, the bias is initialized zero. If it is set to False, this layer will not
have trainable bias parameter. Default: None. have trainable bias parameter. Default: None.
track_running_stats(bool, optional): Whether to compute global stats, which including running mean and
running variance. Default: True.
Shapes: Shapes:
input: Tensor that the dimension from 2 to 5. input: Tensor that the dimension from 2 to 5.
...@@ -1055,11 +1012,10 @@ class SyncBatchNorm(_BatchNormBase): ...@@ -1055,11 +1012,10 @@ class SyncBatchNorm(_BatchNormBase):
weight_attr=None, weight_attr=None,
bias_attr=None, bias_attr=None,
data_format='NCHW', data_format='NCHW',
track_running_stats=True,
name=None): name=None):
super(SyncBatchNorm, super(SyncBatchNorm,
self).__init__(num_features, momentum, epsilon, weight_attr, self).__init__(num_features, momentum, epsilon, weight_attr,
bias_attr, data_format, track_running_stats, name) bias_attr, data_format, name)
def forward(self, x): def forward(self, x):
# create output # create output
...@@ -1147,10 +1103,10 @@ class SyncBatchNorm(_BatchNormBase): ...@@ -1147,10 +1103,10 @@ class SyncBatchNorm(_BatchNormBase):
""" """
layer_output = layer layer_output = layer
if isinstance(layer, _BatchNormBase): if isinstance(layer, _BatchNormBase):
layer_output = SyncBatchNorm( layer_output = SyncBatchNorm(layer._num_features, layer._momentum,
layer._num_features, layer._momentum, layer._epsilon, layer._epsilon, layer._weight_attr,
layer._weight_attr, layer._bias_attr, layer._data_format, layer._bias_attr, layer._data_format,
layer._track_running_stats, layer._name) layer._name)
if layer._weight_attr != False and layer._bias_attr != False: if layer._weight_attr != False and layer._bias_attr != False:
with no_grad(): with no_grad():
......
...@@ -35,7 +35,7 @@ __all__ = [ ...@@ -35,7 +35,7 @@ __all__ = [
class AvgPool1D(layers.Layer): class AvgPool1D(layers.Layer):
""" """
This operation applies a 1D average pooling over an input signal composed This operation applies a 1D average pooling over an input signal composed
of several input planes, based on the input, output_size, return_indices parameters. of several input planes, based on the input, output_size, return_mask parameters.
Input(X) and output(Out) are in NCL format, where N is batch Input(X) and output(Out) are in NCL format, where N is batch
size, C is the number of channels, L is the length of the feature. size, C is the number of channels, L is the length of the feature.
The output tensor shape will be [N, C, output_size]. The output tensor shape will be [N, C, output_size].
...@@ -61,7 +61,7 @@ class AvgPool1D(layers.Layer): ...@@ -61,7 +61,7 @@ class AvgPool1D(layers.Layer):
4. A list[int] or tuple(int) whose length is 2. It has the form [pad_before, pad_after]. 4. A list[int] or tuple(int) whose length is 2. It has the form [pad_before, pad_after].
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
count_include_pad (bool): Whether to exclude padding points in average pooling exclusive (bool): Whether to exclude padding points in average pooling
mode, default is `True`. mode, default is `True`.
ceil_mode (bool): ${ceil_mode_comment}Whether to use the ceil function to calculate output height and width. ceil_mode (bool): ${ceil_mode_comment}Whether to use the ceil function to calculate output height and width.
If it is set to False, the floor function will be used. The default value is False. If it is set to False, the floor function will be used. The default value is False.
...@@ -103,7 +103,7 @@ class AvgPool1D(layers.Layer): ...@@ -103,7 +103,7 @@ class AvgPool1D(layers.Layer):
kernel_size, kernel_size,
stride=None, stride=None,
padding=0, padding=0,
count_include_pad=True, exclusive=True,
ceil_mode=False, ceil_mode=False,
name=None): name=None):
super(AvgPool1D, self).__init__() super(AvgPool1D, self).__init__()
...@@ -111,12 +111,12 @@ class AvgPool1D(layers.Layer): ...@@ -111,12 +111,12 @@ class AvgPool1D(layers.Layer):
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.count_include_pad = count_include_pad self.exclusive = exclusive
self.name = name self.name = name
def forward(self, x): def forward(self, x):
out = F.avg_pool1d(x, self.kernel_size, self.stride, self.padding, out = F.avg_pool1d(x, self.kernel_size, self.stride, self.padding,
self.count_include_pad, self.ceil_mode, self.name) self.exclusive, self.ceil_mode, self.name)
return out return out
...@@ -156,7 +156,7 @@ class AvgPool2D(layers.Layer): ...@@ -156,7 +156,7 @@ class AvgPool2D(layers.Layer):
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
ceil_mode (bool): when True, will use `ceil` instead of `floor` to compute the output shape ceil_mode (bool): when True, will use `ceil` instead of `floor` to compute the output shape
count_include_pad (bool): Whether to exclude padding points in average pooling exclusive (bool): Whether to exclude padding points in average pooling
mode, default is `true`. mode, default is `true`.
divisor_override (float): if specified, it will be used as divisor, otherwise kernel_size will be used. Default None. divisor_override (float): if specified, it will be used as divisor, otherwise kernel_size will be used. Default None.
data_format (string): The data format of the input and output data. An optional string from: `"NCHW"`, `"NDHW"`. data_format (string): The data format of the input and output data. An optional string from: `"NCHW"`, `"NDHW"`.
...@@ -197,7 +197,7 @@ class AvgPool2D(layers.Layer): ...@@ -197,7 +197,7 @@ class AvgPool2D(layers.Layer):
stride=None, stride=None,
padding=0, padding=0,
ceil_mode=False, ceil_mode=False,
count_include_pad=True, exclusive=True,
divisor_override=None, divisor_override=None,
data_format="NCHW", data_format="NCHW",
name=None): name=None):
...@@ -206,7 +206,7 @@ class AvgPool2D(layers.Layer): ...@@ -206,7 +206,7 @@ class AvgPool2D(layers.Layer):
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.count_include_pad = count_include_pad self.exclusive = exclusive
self.divisor = divisor_override self.divisor = divisor_override
self.data_format = data_format self.data_format = data_format
self.name = name self.name = name
...@@ -218,7 +218,7 @@ class AvgPool2D(layers.Layer): ...@@ -218,7 +218,7 @@ class AvgPool2D(layers.Layer):
stride=self.stride, stride=self.stride,
padding=self.padding, padding=self.padding,
ceil_mode=self.ceil_mode, ceil_mode=self.ceil_mode,
count_include_pad=self.count_include_pad, exclusive=self.exclusive,
divisor_override=self.divisor, divisor_override=self.divisor,
data_format=self.data_format, data_format=self.data_format,
name=self.name) name=self.name)
...@@ -247,7 +247,7 @@ class AvgPool3D(layers.Layer): ...@@ -247,7 +247,7 @@ class AvgPool3D(layers.Layer):
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
ceil_mode (bool): ${ceil_mode_comment} ceil_mode (bool): ${ceil_mode_comment}
count_include_pad (bool): Whether to exclude padding points in average pooling exclusive (bool): Whether to exclude padding points in average pooling
mode, default is True. mode, default is True.
divisor_override (int|float) if specified, it will be used as divisor, otherwise kernel_size will be used. Default None. divisor_override (int|float) if specified, it will be used as divisor, otherwise kernel_size will be used. Default None.
data_format (string): The data format of the input and output data. An optional string from: `"NCDHW"`, `"NDHWC"`. data_format (string): The data format of the input and output data. An optional string from: `"NCDHW"`, `"NDHWC"`.
...@@ -289,7 +289,7 @@ class AvgPool3D(layers.Layer): ...@@ -289,7 +289,7 @@ class AvgPool3D(layers.Layer):
stride, stride,
padding=0, padding=0,
ceil_mode=False, ceil_mode=False,
count_include_pad=True, exclusive=True,
divisor_override=None, divisor_override=None,
data_format="NCDHW", data_format="NCDHW",
name=None): name=None):
...@@ -298,7 +298,7 @@ class AvgPool3D(layers.Layer): ...@@ -298,7 +298,7 @@ class AvgPool3D(layers.Layer):
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.count_include_pad = count_include_pad self.exclusive = exclusive
self.divisor = divisor_override self.divisor = divisor_override
self.data_format = data_format self.data_format = data_format
self.name = name self.name = name
...@@ -310,7 +310,7 @@ class AvgPool3D(layers.Layer): ...@@ -310,7 +310,7 @@ class AvgPool3D(layers.Layer):
stride=self.stride, stride=self.stride,
padding=self.padding, padding=self.padding,
ceil_mode=self.ceil_mode, ceil_mode=self.ceil_mode,
count_include_pad=self.count_include_pad, exclusive=self.exclusive,
divisor_override=self.divisor, divisor_override=self.divisor,
data_format=self.data_format, data_format=self.data_format,
name=self.name) name=self.name)
...@@ -319,7 +319,7 @@ class AvgPool3D(layers.Layer): ...@@ -319,7 +319,7 @@ class AvgPool3D(layers.Layer):
class MaxPool1D(layers.Layer): class MaxPool1D(layers.Layer):
""" """
Applies a 1D max pooling over an input signal composed of several input planes based Applies a 1D max pooling over an input signal composed of several input planes based
on the input, output_size, return_indices parameters. on the input, output_size, return_mask parameters.
Input(X) and output(Out) are in NCL format, where N is batch Input(X) and output(Out) are in NCL format, where N is batch
size, C is the number of channels, L is the length of the feature. size, C is the number of channels, L is the length of the feature.
...@@ -343,7 +343,7 @@ class MaxPool1D(layers.Layer): ...@@ -343,7 +343,7 @@ class MaxPool1D(layers.Layer):
4. A list[int] or tuple(int) whose length is 2. It has the form [pad_before, pad_after]. 4. A list[int] or tuple(int) whose length is 2. It has the form [pad_before, pad_after].
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
return_indices (bool): Whether return the max indices along with the outputs. default is `False`. return_mask (bool): Whether return the max indices along with the outputs. default is `False`.
ceil_mode (bool): Whether to use the ceil function to calculate output height and width. False is the default. ceil_mode (bool): Whether to use the ceil function to calculate output height and width. False is the default.
If it is set to False, the floor function will be used. Default False. If it is set to False, the floor function will be used. Default False.
name(str, optional): For detailed information, please refer name(str, optional): For detailed information, please refer
...@@ -377,7 +377,7 @@ class MaxPool1D(layers.Layer): ...@@ -377,7 +377,7 @@ class MaxPool1D(layers.Layer):
pool_out = MaxPool1D(data) pool_out = MaxPool1D(data)
# pool_out shape: [1, 3, 16] # pool_out shape: [1, 3, 16]
MaxPool1D = nn.MaxPool1D(kernel_size=2, stride=2, padding=0, return_indices=True) MaxPool1D = nn.MaxPool1D(kernel_size=2, stride=2, padding=0, return_mask=True)
pool_out, indices = MaxPool1D(data) pool_out, indices = MaxPool1D(data)
# pool_out shape: [1, 3, 16], indices shape: [1, 3, 16] # pool_out shape: [1, 3, 16], indices shape: [1, 3, 16]
...@@ -387,7 +387,7 @@ class MaxPool1D(layers.Layer): ...@@ -387,7 +387,7 @@ class MaxPool1D(layers.Layer):
kernel_size, kernel_size,
stride=None, stride=None,
padding=0, padding=0,
return_indices=False, return_mask=False,
ceil_mode=False, ceil_mode=False,
name=None): name=None):
super(MaxPool1D, self).__init__() super(MaxPool1D, self).__init__()
...@@ -395,12 +395,12 @@ class MaxPool1D(layers.Layer): ...@@ -395,12 +395,12 @@ class MaxPool1D(layers.Layer):
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.return_indices = return_indices self.return_mask = return_mask
self.name = name self.name = name
def forward(self, input): def forward(self, input):
out = F.max_pool1d(input, self.kernel_size, self.stride, self.padding, out = F.max_pool1d(input, self.kernel_size, self.stride, self.padding,
self.return_indices, self.ceil_mode, self.name) self.return_mask, self.ceil_mode, self.name)
return out return out
...@@ -440,7 +440,7 @@ class MaxPool2D(layers.Layer): ...@@ -440,7 +440,7 @@ class MaxPool2D(layers.Layer):
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
ceil_mode (bool): when True, will use `ceil` instead of `floor` to compute the output shape ceil_mode (bool): when True, will use `ceil` instead of `floor` to compute the output shape
return_indices (bool): Whether to return the max indices along with the outputs. return_mask (bool): Whether to return the max indices along with the outputs.
data_format (string): The data format of the input and output data. An optional string from: `"NCHW"`, `"NDHW"`. data_format (string): The data format of the input and output data. An optional string from: `"NCHW"`, `"NDHW"`.
The default is `"NCHW"`. When it is `"NCHW"`, the data is stored in the order of: The default is `"NCHW"`. When it is `"NCHW"`, the data is stored in the order of:
`[batch_size, input_channels, input_height, input_width]`. `[batch_size, input_channels, input_height, input_width]`.
...@@ -473,8 +473,8 @@ class MaxPool2D(layers.Layer): ...@@ -473,8 +473,8 @@ class MaxPool2D(layers.Layer):
output = MaxPool2D(input) output = MaxPool2D(input)
# output.shape [1, 3, 16, 16] # output.shape [1, 3, 16, 16]
# for return_indices=True # for return_mask=True
MaxPool2D = nn.MaxPool2D(kernel_size=2,stride=2, padding=0, return_indices=True) MaxPool2D = nn.MaxPool2D(kernel_size=2, stride=2, padding=0, return_mask=True)
output, max_indices = MaxPool2D(input) output, max_indices = MaxPool2D(input)
# output.shape [1, 3, 16, 16], max_indices.shape [1, 3, 16, 16], # output.shape [1, 3, 16, 16], max_indices.shape [1, 3, 16, 16],
""" """
...@@ -483,7 +483,7 @@ class MaxPool2D(layers.Layer): ...@@ -483,7 +483,7 @@ class MaxPool2D(layers.Layer):
kernel_size, kernel_size,
stride=None, stride=None,
padding=0, padding=0,
return_indices=False, return_mask=False,
ceil_mode=False, ceil_mode=False,
data_format="NCHW", data_format="NCHW",
name=None): name=None):
...@@ -491,7 +491,7 @@ class MaxPool2D(layers.Layer): ...@@ -491,7 +491,7 @@ class MaxPool2D(layers.Layer):
self.ksize = kernel_size self.ksize = kernel_size
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.return_indices = return_indices self.return_mask = return_mask
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.data_format = data_format self.data_format = data_format
self.name = name self.name = name
...@@ -502,7 +502,7 @@ class MaxPool2D(layers.Layer): ...@@ -502,7 +502,7 @@ class MaxPool2D(layers.Layer):
kernel_size=self.ksize, kernel_size=self.ksize,
stride=self.stride, stride=self.stride,
padding=self.padding, padding=self.padding,
return_indices=self.return_indices, return_mask=self.return_mask,
data_format=self.data_format, data_format=self.data_format,
name=self.name) name=self.name)
...@@ -530,7 +530,7 @@ class MaxPool3D(layers.Layer): ...@@ -530,7 +530,7 @@ class MaxPool3D(layers.Layer):
5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0). 5. A list or tuple of pairs of integers. It has the form [[pad_before, pad_after], [pad_before, pad_after], ...]. Note that, the batch dimension and channel dimension should be [0,0] or (0,0).
The default value is 0. The default value is 0.
ceil_mode (bool): ${ceil_mode_comment} ceil_mode (bool): ${ceil_mode_comment}
return_indices (bool): Whether to return the max indices along with the outputs. return_mask (bool): Whether to return the max indices along with the outputs.
data_format (string): The data format of the input and output data. An optional string from: `"NCDHW"`, `"NDHWC"`. data_format (string): The data format of the input and output data. An optional string from: `"NCDHW"`, `"NDHWC"`.
The default is `"NCDHW"`. When it is `"NCDHW"`, the data is stored in the order of: The default is `"NCDHW"`. When it is `"NCDHW"`, the data is stored in the order of:
`[batch_size, input_channels, input_depth, input_height, input_width]`. `[batch_size, input_channels, input_depth, input_height, input_width]`.
...@@ -564,8 +564,8 @@ class MaxPool3D(layers.Layer): ...@@ -564,8 +564,8 @@ class MaxPool3D(layers.Layer):
output = MaxPool3D(input) output = MaxPool3D(input)
# output.shape [1, 2, 3, 16, 16] # output.shape [1, 2, 3, 16, 16]
# for return_indices=True # for return_mask=True
MaxPool3D = nn.MaxPool3D(kernel_size=2,stride=2, padding=0, return_indices=True) MaxPool3D = nn.MaxPool3D(kernel_size=2, stride=2, padding=0, return_mask=True)
output, max_indices = MaxPool3D(input) output, max_indices = MaxPool3D(input)
# output.shape [1, 2, 3, 16, 16], max_indices.shape [1, 2, 3, 16, 16], # output.shape [1, 2, 3, 16, 16], max_indices.shape [1, 2, 3, 16, 16],
""" """
...@@ -574,7 +574,7 @@ class MaxPool3D(layers.Layer): ...@@ -574,7 +574,7 @@ class MaxPool3D(layers.Layer):
kernel_size, kernel_size,
stride, stride,
padding, padding,
return_indices=False, return_mask=False,
ceil_mode=False, ceil_mode=False,
data_format="NCDHW", data_format="NCDHW",
name=None): name=None):
...@@ -582,7 +582,7 @@ class MaxPool3D(layers.Layer): ...@@ -582,7 +582,7 @@ class MaxPool3D(layers.Layer):
self.ksize = kernel_size self.ksize = kernel_size
self.stride = stride self.stride = stride
self.padding = padding self.padding = padding
self.return_indices = return_indices self.return_mask = return_mask
self.ceil_mode = ceil_mode self.ceil_mode = ceil_mode
self.data_format = data_format self.data_format = data_format
self.name = name self.name = name
...@@ -593,7 +593,7 @@ class MaxPool3D(layers.Layer): ...@@ -593,7 +593,7 @@ class MaxPool3D(layers.Layer):
kernel_size=self.ksize, kernel_size=self.ksize,
stride=self.stride, stride=self.stride,
padding=self.padding, padding=self.padding,
return_indices=self.return_indices, return_mask=self.return_mask,
data_format=self.data_format, data_format=self.data_format,
name=self.name) name=self.name)
...@@ -602,7 +602,7 @@ class AdaptiveAvgPool1D(layers.Layer): ...@@ -602,7 +602,7 @@ class AdaptiveAvgPool1D(layers.Layer):
""" """
This operation applies a 1D adaptive average pooling over an input signal composed This operation applies a 1D adaptive average pooling over an input signal composed
of several input planes, based on the input, output_size, return_indices parameters. of several input planes, based on the input, output_size, return_mask parameters.
Input(X) and output(Out) are in NCL format, where N is batch Input(X) and output(Out) are in NCL format, where N is batch
size, C is the number of channels, L is the length of the feature. size, C is the number of channels, L is the length of the feature.
The output tensor shape will be [N, C, output_size]. The output tensor shape will be [N, C, output_size].
...@@ -841,7 +841,7 @@ class AdaptiveMaxPool1D(layers.Layer): ...@@ -841,7 +841,7 @@ class AdaptiveMaxPool1D(layers.Layer):
""" """
This operation applies a 1D adaptive max pooling over an input signal composed This operation applies a 1D adaptive max pooling over an input signal composed
of several input planes, based on the input, output_size, return_indices parameters. of several input planes, based on the input, output_size, return_mask parameters.
Input(X) and output(Out) are in NCL format, where N is batch Input(X) and output(Out) are in NCL format, where N is batch
size, C is the number of channels, L is the length of the feature. size, C is the number of channels, L is the length of the feature.
The output tensor shape will be [N, C, output_size]. The output tensor shape will be [N, C, output_size].
...@@ -859,7 +859,7 @@ class AdaptiveMaxPool1D(layers.Layer): ...@@ -859,7 +859,7 @@ class AdaptiveMaxPool1D(layers.Layer):
Args: Args:
output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list, output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list,
it must contain one int. it must contain one int.
return_indices (bool): If true, the index of max pooling point will be returned along return_mask (bool): If true, the index of max pooling point will be returned along
with outputs. It cannot be set in average pooling type. Default False. with outputs. It cannot be set in average pooling type. Default False.
name(str, optional): For detailed information, please refer name(str, optional): For detailed information, please refer
to :ref:`api_guide_Name`. Usually name is no need to set and to :ref:`api_guide_Name`. Usually name is no need to set and
...@@ -898,22 +898,22 @@ class AdaptiveMaxPool1D(layers.Layer): ...@@ -898,22 +898,22 @@ class AdaptiveMaxPool1D(layers.Layer):
pool_out = AdaptiveMaxPool1D(data) pool_out = AdaptiveMaxPool1D(data)
# pool_out shape: [1, 3, 16] # pool_out shape: [1, 3, 16]
# for return_indices = true # for return_mask = true
AdaptiveMaxPool1D = nn.AdaptiveMaxPool1D(output_size=16, return_indices=True) AdaptiveMaxPool1D = nn.AdaptiveMaxPool1D(output_size=16, return_mask=True)
pool_out, indices = AdaptiveMaxPool1D(data) pool_out, indices = AdaptiveMaxPool1D(data)
# pool_out shape: [1, 3, 16], indices shape: [1, 3, 16] # pool_out shape: [1, 3, 16], indices shape: [1, 3, 16]
""" """
def __init__(self, output_size, return_indices=False, name=None): def __init__(self, output_size, return_mask=False, name=None):
super(AdaptiveMaxPool1D, self).__init__() super(AdaptiveMaxPool1D, self).__init__()
self.output_size = output_size self.output_size = output_size
self.return_indices = return_indices self.return_mask = return_mask
self.name = name self.name = name
def forward(self, input): def forward(self, input):
return F.adaptive_max_pool1d(input, self.output_size, return F.adaptive_max_pool1d(input, self.output_size, self.return_mask,
self.return_indices, self.name) self.name)
class AdaptiveMaxPool2D(layers.Layer): class AdaptiveMaxPool2D(layers.Layer):
...@@ -932,7 +932,7 @@ class AdaptiveMaxPool2D(layers.Layer): ...@@ -932,7 +932,7 @@ class AdaptiveMaxPool2D(layers.Layer):
Output(i ,j) &= max(Input[hstart:hend, wstart:wend]) Output(i ,j) &= max(Input[hstart:hend, wstart:wend])
Parameters: Parameters:
output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list, it must contain two element, (H, W). H and W can be either a int, or None which means the size will be the same as that of the input. output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list, it must contain two element, (H, W). H and W can be either a int, or None which means the size will be the same as that of the input.
return_indices (bool): If true, the index of max pooling point will be returned along with outputs. It cannot be set in average pooling type. Default False. return_mask (bool): If true, the index of max pooling point will be returned along with outputs. It cannot be set in average pooling type. Default False.
name(str, optional): For detailed information, please refer name(str, optional): For detailed information, please refer
to :ref:`api_guide_Name`. Usually name is no need to set and to :ref:`api_guide_Name`. Usually name is no need to set and
None by default. None by default.
...@@ -965,21 +965,21 @@ class AdaptiveMaxPool2D(layers.Layer): ...@@ -965,21 +965,21 @@ class AdaptiveMaxPool2D(layers.Layer):
paddle.disable_static() paddle.disable_static()
input_data = np.random.rand(2, 3, 32, 32) input_data = np.random.rand(2, 3, 32, 32)
x = paddle.to_tensor(input_data) x = paddle.to_tensor(input_data)
adaptive_max_pool = paddle.nn.AdaptiveMaxPool2D(output_size=3, return_indices=True) adaptive_max_pool = paddle.nn.AdaptiveMaxPool2D(output_size=3, return_mask=True)
pool_out, indices = adaptive_max_pool(x = x) pool_out, indices = adaptive_max_pool(x = x)
""" """
def __init__(self, output_size, return_indices=False, name=None): def __init__(self, output_size, return_mask=False, name=None):
super(AdaptiveMaxPool2D, self).__init__() super(AdaptiveMaxPool2D, self).__init__()
self._output_size = output_size self._output_size = output_size
self._return_indices = return_indices self._return_mask = return_mask
self._name = name self._name = name
def forward(self, x): def forward(self, x):
return F.adaptive_max_pool2d( return F.adaptive_max_pool2d(
x, x,
output_size=self._output_size, output_size=self._output_size,
return_indices=self._return_indices, return_mask=self._return_mask,
name=self._name) name=self._name)
...@@ -1002,7 +1002,7 @@ class AdaptiveMaxPool3D(layers.Layer): ...@@ -1002,7 +1002,7 @@ class AdaptiveMaxPool3D(layers.Layer):
Parameters: Parameters:
output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list, it must contain three elements, (D, H, W). D, H and W can be either a int, or None which means the size will be the same as that of the input. output_size (int|list|tuple): The pool kernel size. If pool kernel size is a tuple or list, it must contain three elements, (D, H, W). D, H and W can be either a int, or None which means the size will be the same as that of the input.
return_indices (bool): If true, the index of max pooling point will be returned along with outputs. Default False. return_mask (bool): If true, the index of max pooling point will be returned along with outputs. Default False.
name(str, optional): For detailed information, please refer name(str, optional): For detailed information, please refer
to :ref:`api_guide_Name`. Usually name is no need to set and to :ref:`api_guide_Name`. Usually name is no need to set and
None by default. None by default.
...@@ -1040,21 +1040,21 @@ class AdaptiveMaxPool3D(layers.Layer): ...@@ -1040,21 +1040,21 @@ class AdaptiveMaxPool3D(layers.Layer):
pool = paddle.nn.AdaptiveMaxPool3D(output_size=4) pool = paddle.nn.AdaptiveMaxPool3D(output_size=4)
out = pool(x) out = pool(x)
# out shape: [2, 3, 4, 4, 4] # out shape: [2, 3, 4, 4, 4]
pool = paddle.nn.AdaptiveMaxPool3D(output_size=3, return_indices=True) pool = paddle.nn.AdaptiveMaxPool3D(output_size=3, return_mask=True)
out, indices = pool(x) out, indices = pool(x)
# out shape: [2, 3, 4, 4, 4], indices shape: [2, 3, 4, 4, 4] # out shape: [2, 3, 4, 4, 4], indices shape: [2, 3, 4, 4, 4]
""" """
def __init__(self, output_size, return_indices=False, name=None): def __init__(self, output_size, return_mask=False, name=None):
super(AdaptiveMaxPool3D, self).__init__() super(AdaptiveMaxPool3D, self).__init__()
self._output_size = output_size self._output_size = output_size
self._return_indices = return_indices self._return_mask = return_mask
self._name = name self._name = name
def forward(self, x): def forward(self, x):
return F.adaptive_max_pool3d( return F.adaptive_max_pool3d(
x, x,
output_size=self._output_size, output_size=self._output_size,
return_indices=self._return_indices, return_mask=self._return_mask,
name=self._name) name=self._name)
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册