tf.nn.selu

tf.nn.selu(
    features,
    name=None
)

Defined in generated file: tensorflow/python/ops/gen_nn_ops.py.

See the guide: Neural Network > Activation Functions

Computes scaled exponential linear: scale * alpha * (exp(features) - 1)

if < 0, scale * features otherwise.

To be used together with initializer = tf.variance_scaling_initializer(factor=1.0, mode='FAN_IN'). For correct dropout, use tf.contrib.nn.alpha_dropout.

See Self-Normalizing Neural Networks

Args:

  • features: A Tensor. Must be one of the following types: half, bfloat16, float32, float64.
  • name: A name for the operation (optional).

Returns:

A Tensor. Has the same type as features.