tf.keras.activations.relu(
x,
alpha=0.0,
max_value=None,
threshold=0
)
Defined in tensorflow/python/keras/activations.py
.
Rectified Linear Unit.
With default values, it returns element-wise max(x, 0)
.
Otherwise, it follows:
f(x) = max_value
for x >= max_value
,
f(x) = x
for threshold <= x < max_value
,
f(x) = alpha * (x - threshold)
otherwise.
Arguments:
x
: A tensor or variable.alpha
: A scalar, slope of negative section (default=0.
).max_value
: float. Saturation threshold.threshold
: float. Threshold value for thresholded activation.
Returns:
A tensor.