Initializer¶
Initializer¶
- 
class 
paddle.v2.fluid.initializer.Initializer Base class for variable initializers
Defines the common interface of variable initializers. They add operations to the init program that are used to initialize variables. Users should not use this class directly, but need to use one of its implementations.
ConstantInitializer¶
- 
class 
paddle.v2.fluid.initializer.ConstantInitializer(value=0.0) Implements the constant initializer
UniformInitializer¶
- 
class 
paddle.v2.fluid.initializer.UniformInitializer(low=-1.0, high=1.0, seed=0) Implements the random uniform distribution initializer
NormalInitializer¶
- 
class 
paddle.v2.fluid.initializer.NormalInitializer(loc=0.0, scale=1.0, seed=0) Implements the random Normal(Gaussian) distribution initializer
XavierInitializer¶
- 
class 
paddle.v2.fluid.initializer.XavierInitializer(uniform=True, fan_in=None, fan_out=None, seed=0) Implements the Xavier initializer
This class implements the Xavier weight initializer from the paper Understanding the difficulty of training deep feedforward neural networks[1] by Xavier Glorot and Yoshua Bengio.
This initializer is designed to keep the scale of the gradients approximately same in all the layers. In case of Uniform distribution, the range is [-x, x], where x = sqrt(6 / (fan_in + fan_out)). In case of Normal distribution, the mean is 0 and the standard deviation is sqrt(2/ (fan_in + fan_out)).
References
- [1] Understanding the difficulty of training deep feedforward neural
 - networks. International conference on artificial intelligence and statistics. (http://proceedings.mlr.press/v9/glorot10a.html)
 
MSRAInitializer¶
- 
class 
paddle.v2.fluid.initializer.MSRAInitializer(uniform=True, fan_in=None, seed=0) Implements the MSRA initializer a.k.a. Kaiming Initializer
This class implements the weight initialization from the paper Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification[1] by Kaiming He, Xiangyu Zhang, Shaoqing Ren and Jian Sun. This is a robust initialization method that particularly considers the rectifier nonlinearities. In case of Uniform distribution, the range is [-x, x], where x = sqrt(6 / fan_in). In case of Normal distribution, the mean is 0 and the standard deviation is sqrt(2/ fan_in).
References
- [1] Delving Deep into Rectifiers: Surpassing Human-Level Performance
 - on ImageNet Classification (https://arxiv.org/abs/1502.01852)
 
