tf.keras.activations.relu

tf.keras.activations.relu(
    x,
    alpha=0.0,
    max_value=None,
    threshold=0
)

Defined in tensorflow/python/keras/activations.py.

Rectified Linear Unit.

With default values, it returns element-wise max(x, 0).

Otherwise, it follows: f(x) = max_value for x >= max_value, f(x) = x for threshold <= x < max_value, f(x) = alpha * (x - threshold) otherwise.

Arguments:

  • x: A tensor or variable.
  • alpha: A scalar, slope of negative section (default=0.).
  • max_value: float. Saturation threshold.
  • threshold: float. Threshold value for thresholded activation.

Returns:

A tensor.