tf.keras.activations.elu

Exponential linear unit.

Aliases:

  • tf.compat.v1.keras.activations.elu
  • tf.compat.v2.keras.activations.elu
  • tf.keras.activations.elu
tf.keras.activations.elu(
    x,
    alpha=1.0
)

Defined in python/keras/activations.py.

Arguments:

  • x: Input tensor.
  • alpha: A scalar, slope of negative section.

Returns:

The exponential linear activation: x if x > 0 and alpha * (exp(x)-1) if x < 0.

Reference: