- Sort Score
- Result 10 results
- Languages All
Results 1 - 1 of 1 for LogSoftmax (0.07 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/lower_tf.td
// computes loss and backprop of the loss with respect to 'features'. // // Softmax cross entropy loss is defined as follows: // // loss = Sum(-labels * Log(Exp(features) / Sum(Exp(features))) // loss = Sum(-labels * LogSoftmax(features)) // // Computing gradient of the loss with respect to features gives us, // // backprop = (Exp(features) / Sum(Exp(features))) - labels // backprop = Softmax(features) - labels //
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 04 13:30:42 UTC 2024 - 24.7K bytes - Viewed (0)