- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for Peckover (0.22 sec)
-
tensorflow/c/experimental/gradients/nn_grad.cc
* * dA = U * dbias = reduceSum(U, dims = channel_dim) */ AbstractTensorHandle* upstream_grad = grad_outputs[0]; DCHECK(upstream_grad); // Recover data format from forward pass for gradient. std::string data_format; TF_RETURN_IF_ERROR(forward_attrs_.Get("data_format", &data_format)); // Grad for A grad_inputs[0] = upstream_grad;
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
per_device_output_tensors.push_back(device_thread->Join(status)); // We will run every Join even if there are bad statuses in case the user // wants to recover and continue running ops on the parallel device (which // would otherwise deadlock). if (TF_GetCode(status) != TF_OK && (first_bad_status == nullptr
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/gradients.cc
// and not for correctness. The only downside of keeping this 1 seems to be // that the gradient accumulation is unbounded and we will never // aggressively aggregate accumulated gradients to recover memory. // Revisit and fix. return 1; } // Consumes references to the tensors in the gradient_tensors list and returns // a tensor with the result. AbstractTensorHandle* TapeVSpace::AggregateGradients(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0)