|View source on GitHub|
Adds an Absolute Difference loss to the training procedure.
tf.losses.absolute_difference( labels, predictions, weights=1.0, scope=None, loss_collection=tf.GraphKeys.LOSSES, reduction=Reduction.SUM_BY_NONZERO_WEIGHTS )
weights acts as a coefficient for the loss. If a scalar is provided, then
the loss is simply scaled by the given value. If
weights is a
[batch_size], then the total loss for each sample of the batch is
rescaled by the corresponding element in the
weights vector. If the shape of
weights matches the shape of
predictions, then the loss of each
measurable element of
predictions is scaled by the corresponding value of
labels: The ground truth output tensor, same dimensions as 'predictions'.
predictions: The predicted outputs.
Tensorwhose rank is either 0, or the same rank as
labels, and must be broadcastable to
labels(i.e., all dimensions must be either
1, or the same as the corresponding
scope: The scope for the operations performed in computing the loss.
loss_collection: collection to which this loss will be added.
reduction: Type of reduction to apply to loss.
Weighted loss float
NONE, this has the same
labels; otherwise, it is scalar.
ValueError: If the shape of
predictionsdoesn't match that of
labelsor if the shape of
weightsis invalid or if
loss_collection argument is ignored when executing eagerly. Consider
holding on to the return value or collecting losses via a