View source on GitHub |
Compute the Leaky ReLU activation function.
tf.nn.leaky_relu(
features, alpha=0.2, name=None
)
features
: A Tensor
representing preactivation values. Must be one of
the following types: float16
, float32
, float64
, int32
, int64
.alpha
: Slope of the activation function at x < 0.name
: A name for the operation (optional).The activation value.