@@ -591,15 +591,15 @@ class GradScaler(AmpScaler):
...
@@ -591,15 +591,15 @@ class GradScaler(AmpScaler):
Args:
Args:
enable(bool, optional): Enable loss scaling or not. Default is True.
enable(bool, optional): Enable loss scaling or not. Default is True.
init_loss_scaling (float, optional): The initial loss scaling factor. Default is 2**15.
init_loss_scaling (float, optional): The initial loss scaling factor. Default is 65536.0.
incr_ratio(float, optional): The multiplier to use when increasing the loss
incr_ratio(float, optional): The multiplier to use when increasing the loss
scaling. Default is 2.0.
scaling. Default is 2.0.
decr_ratio(float, optional): The less-than-one-multiplier to use when decreasing
decr_ratio(float, optional): The less-than-one-multiplier to use when decreasing
the loss scaling. Default is 0.5.
the loss scaling. Default is 0.5.
incr_every_n_steps(int, optional): Increases loss scaling every n consecutive
incr_every_n_steps(int, optional): Increases loss scaling every n consecutive
steps with finite gradients. Default is 1000.
steps with finite gradients. Default is 2000.
decr_every_n_nan_or_inf(int, optional): Decreases loss scaling every n
decr_every_n_nan_or_inf(int, optional): Decreases loss scaling every n
accumulated steps with nan or inf gradients. Default is 2.
accumulated steps with nan or inf gradients. Default is 1.
use_dynamic_loss_scaling(bool, optional): Whether to use dynamic loss scaling. If False, fixed loss_scaling is used. If True, the loss scaling is updated dynamicly. Default is True.
use_dynamic_loss_scaling(bool, optional): Whether to use dynamic loss scaling. If False, fixed loss_scaling is used. If True, the loss scaling is updated dynamicly. Default is True.
Returns:
Returns:
An GradScaler object.
An GradScaler object.
...
@@ -628,11 +628,11 @@ class GradScaler(AmpScaler):
...
@@ -628,11 +628,11 @@ class GradScaler(AmpScaler):