View source on GitHub |
Softsign activation function, softsign(x) = x / (abs(x) + 1)
.
tf.keras.activations.softsign(
x
)
Example Usage:
a = tf.constant([-1.0, 0.0, 1.0], dtype = tf.float32)
b = tf.keras.activations.softsign(a)
b.numpy()
array([-0.5, 0. , 0.5], dtype=float32)
Args | |
---|---|
x
|
Input tensor. |
Returns | |
---|---|
The softsign activation: x / (abs(x) + 1) .
|