tf.mixed_precision.experimental.FixedLossScale

View source on GitHub

Loss scale with a fixed value.

Inherits From: LossScale

tf.mixed_precision.experimental.FixedLossScale(
    loss_scale_value
)

The loss scale is not updated for the lifetime of instances of this class. A given instance of this class always returns the same number when called.

Args:

Raises:

Methods

__call__

View source

__call__()

Returns the current loss scale as a scalar float32 tensor.

from_config

View source

@classmethod
from_config(
    config
)

Creates the LossScale from its config.

get_config

View source

get_config()

Returns the config of this loss scale.

update

View source

update(
    grads
)

Updates the value of the loss scale.

The loss scale will be potentially updated, based on the value of grads. The tensor returned by calling this class is only updated when this function is evaluated.

In eager mode, this directly updates the loss scale, so that calling __call__ will return the newly updated loss scale. In graph mode, this returns an op that, when evaluated, updates the loss scale.

This function also returns a should_apply_gradients bool. If False, gradients should not be applied to the variables that step, as nonfinite gradients were found, and the loss scale has been be updated to reduce the chance of finding nonfinite gradients in the next step. Some loss scale classes will always return True, as they cannot adjust themselves in response to nonfinite gradients.

When a DistributionStrategy is used, this function may only be called in a cross-replica context.

Args:

Returns: