Measures the probability error in discrete classification tasks in which the
classes are mutually exclusive (each entry is in exactly one class). For
example, each CIFAR-10 image is labeled with one and only one label: an image
can be a dog or a truck, but not both.
If using exclusive labels (wherein one and only
one class is true at a time), see sparse_softmax_cross_entropy_with_logits.
A common use case is to have logits and labels of shape
[batch_size, num_classes], but higher dimensions are supported, with
the axis argument specifying the class dimension.
logits and labels must have the same dtype (either float16, float32,
or float64).
Backpropagation will happen into both logits and labels. To disallow
backpropagation into labels, pass label tensors through tf.stop_gradient
before feeding it to this function.
Note that to avoid confusion, it is required to pass only named arguments to
this function.
Args
labels
Each vector along the class dimension should hold a valid
probability distribution e.g. for the case in which labels are of shape
[batch_size, num_classes], each row of labels[i] must be a valid
probability distribution.
logits
Unscaled log probabilities.
axis
The class dimension. Defaulted to -1 which is the last dimension.
name
A name for the operation (optional).
dim
Deprecated alias for axis.
Returns
A Tensor that contains the softmax cross entropy loss. Its type is the
same as logits and its shape is the same as labels except that it does
not have the last dimension of labels.
[[["Easy to understand","easyToUnderstand","thumb-up"],["Solved my problem","solvedMyProblem","thumb-up"],["Other","otherUp","thumb-up"]],[["Missing the information I need","missingTheInformationINeed","thumb-down"],["Too complicated / too many steps","tooComplicatedTooManySteps","thumb-down"],["Out of date","outOfDate","thumb-down"],["Samples / code issue","samplesCodeIssue","thumb-down"],["Other","otherDown","thumb-down"]],["Last updated 2020-10-01 UTC."],[],[],null,["# tf.compat.v1.nn.softmax_cross_entropy_with_logits_v2\n\n\u003cbr /\u003e\n\n|---------------------------------------------------------------------------------------------------------------------------|\n| [View source on GitHub](https://github.com/tensorflow/tensorflow/blob/v2.0.0/tensorflow/python/ops/nn_ops.py#L3108-L3224) |\n\nComputes softmax cross entropy between `logits` and `labels`. (deprecated arguments) \n\n tf.compat.v1.nn.softmax_cross_entropy_with_logits_v2(\n labels, logits, axis=None, name=None, dim=None\n )\n\n| **Warning:** SOME ARGUMENTS ARE DEPRECATED: `(dim)`. They will be removed in a future version. Instructions for updating: dim is deprecated, use axis instead\n\nMeasures the probability error in discrete classification tasks in which the\nclasses are mutually exclusive (each entry is in exactly one class). For\nexample, each CIFAR-10 image is labeled with one and only one label: an image\ncan be a dog or a truck, but not both.\n| **Note:** While the classes are mutually exclusive, their probabilities need not be. All that is required is that each row of `labels` is a valid probability distribution. If they are not, the computation of the gradient will be incorrect.\n\nIf using exclusive `labels` (wherein one and only\none class is true at a time), see `sparse_softmax_cross_entropy_with_logits`.\n| **Warning:** This op expects unscaled logits, since it performs a `softmax` on `logits` internally for efficiency. Do not call this op with the output of `softmax`, as it will produce incorrect results.\n\nA common use case is to have logits and labels of shape\n`[batch_size, num_classes]`, but higher dimensions are supported, with\nthe `axis` argument specifying the class dimension.\n\n`logits` and `labels` must have the same dtype (either `float16`, `float32`,\nor `float64`).\n\nBackpropagation will happen into both `logits` and `labels`. To disallow\nbackpropagation into `labels`, pass label tensors through [`tf.stop_gradient`](../../../../tf/stop_gradient)\nbefore feeding it to this function.\n\n**Note that to avoid confusion, it is required to pass only named arguments to\nthis function.**\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Args ---- ||\n|----------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|\n| `labels` | Each vector along the class dimension should hold a valid probability distribution e.g. for the case in which labels are of shape `[batch_size, num_classes]`, each row of `labels[i]` must be a valid probability distribution. |\n| `logits` | Unscaled log probabilities. |\n| `axis` | The class dimension. Defaulted to -1 which is the last dimension. |\n| `name` | A name for the operation (optional). |\n| `dim` | Deprecated alias for axis. |\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n\u003cbr /\u003e\n\n| Returns ------- ||\n|---|---|\n| A `Tensor` that contains the softmax cross entropy loss. Its type is the same as `logits` and its shape is the same as `labels` except that it does not have the last dimension of `labels`. ||\n\n\u003cbr /\u003e"]]