uniform.py 9.4 KB
Newer Older
1
# Copyright (c) 2021 PaddlePaddle Authors. All Rights Reserved.
2
#
3 4 5
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
6
#
7
#     http://www.apache.org/licenses/LICENSE-2.0
8
#
9 10 11 12 13 14 15
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import numpy as np
16 17

import paddle
18
from paddle import _C_ops
19
from paddle.distribution import distribution
20
from paddle.fluid.data_feeder import check_type, convert_dtype
21
from paddle.fluid.layers import tensor
22
from paddle.framework import in_dynamic_mode
23
from paddle.tensor import random
24 25 26


class Uniform(distribution.Distribution):
27 28 29 30 31 32 33 34
    r"""Uniform distribution with `low` and `high` parameters.

    Mathematical Details

    The probability density function (pdf) is

    .. math::

35
        pdf(x; a, b) = \frac{1}{Z}, \ a <=x <b
36 37 38 39 40 41 42 43 44 45 46 47

    .. math::

        Z = b - a

    In the above equation:

    * :math:`low = a`,
    * :math:`high = b`,
    * :math:`Z`: is the normalizing constant.

    The parameters `low` and `high` must be shaped in a way that supports
I
Infinity_lee 已提交
48 49 50 51 52 53
    `Boardcasting` (e.g., `high - low` is a valid operation).

    Note:
        If you want know more about broadcasting, please refer to `Introduction to Tensor`_ .

        .. _Introduction to Tensor: ../../guides/beginner/tensor_en.html#chapter5-broadcasting-of-tensor
54 55

    Args:
56 57 58 59 60
        low(int|float|list|tuple|numpy.ndarray|Tensor): The lower boundary of
            uniform distribution.The data type is float32 and float64.
        high(int|float|list|tuple|numpy.ndarray|Tensor): The higher boundary
            of uniform distribution.The data type is float32 and float64.
        name (str, optional): For details, please refer to :ref:`api_guide_Name`. Generally, no setting is required. Default: None.
61 62 63 64

    Examples:
        .. code-block:: python

M
megemini 已提交
65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100
            >>> import paddle
            >>> from paddle.distribution import Uniform
            >>> paddle.seed(2023)

            >>> # Without broadcasting, a single uniform distribution [3, 4]:
            >>> u1 = Uniform(low=3.0, high=4.0)
            >>> # 2 distributions [1, 3], [2, 4]
            >>> u2 = Uniform(low=[1.0, 2.0], high=[3.0, 4.0])
            >>> # 4 distributions
            >>> u3 = Uniform(low=[[1.0, 2.0], [3.0, 4.0]],
            ...             high=[[1.5, 2.5], [3.5, 4.5]])
            ...
            >>> # With broadcasting:
            >>> u4 = Uniform(low=3.0, high=[5.0, 6.0, 7.0])

            >>> # Complete example
            >>> value_tensor = paddle.to_tensor([0.8], dtype="float32")

            >>> uniform = Uniform([0.], [2.])

            >>> sample = uniform.sample([2])
            >>> # a random tensor created by uniform distribution with shape: [2, 1]
            >>> entropy = uniform.entropy()
            >>> print(entropy)
            Tensor(shape=[1], dtype=float32, place=Place(cpu), stop_gradient=True,
                [0.69314718])

            >>> lp = uniform.log_prob(value_tensor)
            >>> print(lp)
            Tensor(shape=[1], dtype=float32, place=Place(cpu), stop_gradient=True,
                [-0.69314718])

            >>> p = uniform.probs(value_tensor)
            >>> print(p)
            Tensor(shape=[1], dtype=float32, place=Place(cpu), stop_gradient=True,
                [0.50000000])
101 102 103
    """

    def __init__(self, low, high, name=None):
104
        if not in_dynamic_mode():
105 106 107 108 109 110 111 112 113 114 115 116
            check_type(
                low,
                'low',
                (int, float, np.ndarray, tensor.Variable, list, tuple),
                'Uniform',
            )
            check_type(
                high,
                'high',
                (int, float, np.ndarray, tensor.Variable, list, tuple),
                'Uniform',
            )
117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134

        self.all_arg_is_float = False
        self.batch_size_unknown = False
        self.name = name if name is not None else 'Uniform'
        self.dtype = 'float32'

        if isinstance(low, int):
            low = float(low)
        if isinstance(high, int):
            high = float(high)

        if self._validate_args(low, high):
            self.low = low
            self.high = high
            self.dtype = convert_dtype(low.dtype)
        else:
            if isinstance(low, float) and isinstance(high, float):
                self.all_arg_is_float = True
135 136 137 138
            if isinstance(low, np.ndarray) and str(low.dtype) in [
                'float32',
                'float64',
            ]:
139
                self.dtype = low.dtype
140 141 142 143
            elif isinstance(high, np.ndarray) and str(high.dtype) in [
                'float32',
                'float64',
            ]:
144 145 146 147
                self.dtype = high.dtype
            # pylint: disable=unbalanced-tuple-unpacking
            self.low, self.high = self._to_tensor(low, high)
            if self.dtype != convert_dtype(self.low.dtype):
148 149
                self.low = paddle.cast(self.low, dtype=self.dtype)
                self.high = paddle.cast(self.high, dtype=self.dtype)
150

151
        super().__init__(self.low.shape)
152

153 154 155 156
    def sample(self, shape, seed=0):
        """Generate samples of the specified shape.

        Args:
157 158
            shape (list): 1D `int32`. Shape of the generated samples.
            seed (int): Python integer number.
159 160

        Returns:
161
            Tensor, A tensor with prepended dimensions shape. The data type is float32.
162 163

        """
164
        if not in_dynamic_mode():
165 166 167 168 169
            check_type(shape, 'shape', (list), 'sample')
            check_type(seed, 'seed', (int), 'sample')

        name = self.name + '_sample'
        batch_shape = list((self.low + self.high).shape)
170
        if -1 in batch_shape:
171 172
            output_shape = shape + batch_shape
            zero_tmp = tensor.fill_constant_batch_size_like(
173 174
                self.low + self.high, batch_shape + shape, self.dtype, 0.0
            )
175
            uniform_random_tmp = random.uniform_random_batch_size_like(
176 177 178
                zero_tmp,
                zero_tmp.shape,
                dtype=self.dtype,
179 180 181 182
                min=0.0,
                max=1.0,
                seed=seed,
            )
183 184
            zero_tmp_reshape = paddle.reshape(zero_tmp, output_shape)
            uniform_random_tmp_reshape = paddle.reshape(
185 186 187 188 189
                uniform_random_tmp, output_shape
            )
            output = uniform_random_tmp_reshape * (
                zero_tmp_reshape + self.high - self.low
            )
190
            output = paddle.add(output, self.low, name=name)
191 192 193
            return output
        else:
            output_shape = shape + batch_shape
194
            output = paddle.uniform(
195 196
                output_shape, dtype=self.dtype, min=0.0, max=1.0, seed=seed
            ) * (
197
                paddle.zeros(output_shape, dtype=self.dtype)
198 199
                + (self.high - self.low)
            )
200
            output = paddle.add(output, self.low, name=name)
201
            if self.all_arg_is_float:
202
                return paddle.reshape(output, shape, name=name)
203 204 205 206 207 208 209
            else:
                return output

    def log_prob(self, value):
        """Log probability density/mass function.

        Args:
210
            value (Tensor): The input tensor.
211 212

        Returns:
213
            Tensor, log probability.The data type is same with value.
214 215 216

        """
        value = self._check_values_dtype_in_probs(self.low, value)
217
        if in_dynamic_mode():
218 219 220 221
            # ensure value in [low, high]
            lb_bool = self.low < value
            ub_bool = value < self.high

222 223 224 225 226 227 228
            lb = _C_ops.cast(lb_bool, value.dtype)
            ub = _C_ops.cast(ub_bool, value.dtype)
            return paddle.log(lb * ub) - paddle.log(self.high - self.low)
        else:
            name = self.name + '_log_prob'
            lb_bool = self.low < value
            ub_bool = value < self.high
229 230
            lb = paddle.cast(lb_bool, dtype=value.dtype)
            ub = paddle.cast(ub_bool, dtype=value.dtype)
231 232 233
            return paddle.subtract(
                paddle.log(lb * ub), paddle.log(self.high - self.low), name=name
            )
234 235 236 237 238

    def probs(self, value):
        """Probability density/mass function.

        Args:
239
            value (Tensor): The input tensor.
240 241

        Returns:
242
            Tensor, probability. The data type is same with value.
243 244 245

        """
        value = self._check_values_dtype_in_probs(self.low, value)
246
        if in_dynamic_mode():
247 248
            lb_bool = self.low < value
            ub_bool = value < self.high
249 250 251 252 253 254 255
            lb = _C_ops.cast(lb_bool, value.dtype)
            ub = _C_ops.cast(ub_bool, value.dtype)
            return (lb * ub) / (self.high - self.low)
        else:
            name = self.name + '_probs'
            lb_bool = self.low < value
            ub_bool = value < self.high
256 257
            lb = paddle.cast(lb_bool, dtype=value.dtype)
            ub = paddle.cast(ub_bool, dtype=value.dtype)
258
            return paddle.divide((lb * ub), (self.high - self.low), name=name)
259 260 261 262 263 264 265 266 267 268 269

    def entropy(self):
        r"""Shannon entropy in nats.

        The entropy is

        .. math::

            entropy(low, high) = \\log (high - low)

        Returns:
270
            Tensor, Shannon entropy of uniform distribution.The data type is float32.
271 272 273

        """
        name = self.name + '_entropy'
274
        return paddle.log(self.high - self.low, name=name)