Warning: This project is deprecated. TensorFlow Addons has stopped development, The project will only be providing minimal maintenance releases until May 2024. See the full announcement here or on github.

tfa.activations.mish

Mish: A Self Regularized Non-Monotonic Neural Activation Function.

Computes mish activation:

\[ \mathrm{mish}(x) = x \cdot \tanh(\mathrm{softplus}(x)). \]

See Mish: A Self Regularized Non-Monotonic Neural Activation Function.

Usage:

x = tf.constant([1.0, 0.0, 1.0])
tfa.activations.mish(x)
<tf.Tensor: shape=(3,), dtype=float32, numpy=array([0.865098..., 0.       , 0.865098...], dtype=float32)>

x A Tensor. Must be one of the following types: bfloat16, float16, float32, float64.

A Tensor. Has the same type as x.