- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for removed (0.17 sec)
-
tensorflow/c/eager/parallel_device/parallel_device_remote_test.cc
A. Unique TensorFlower <******@****.***> 1682633218 -0700
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 22:09:57 GMT 2023 - 6.7K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad.cc
* * dA = U * dbias = reduceSum(U, dims = channel_dim) */ AbstractTensorHandle* upstream_grad = grad_outputs[0]; DCHECK(upstream_grad); // Recover data format from forward pass for gradient. std::string data_format; TF_RETURN_IF_ERROR(forward_attrs_.Get("data_format", &data_format)); // Grad for A grad_inputs[0] = upstream_grad;
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/c_api.cc
namespace { #if !defined(IS_MOBILE_PLATFORM) && !defined(IS_SLIM_BUILD) // Creates a placeholder representing an input to the cond or body graph. // TODO(skyewm): remove these from final graph bool CreateInput(const TF_Output& parent_input, TF_Graph* g, const char* name, TF_Output* input, TF_Status* status) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/math_grad_test.cc
immediate_execution_ctx_.get(), {x.get()}, UseFunction())); } TEST_P(CppGradients, TestMatMulGrad) { // TODO(vnvo2409): Figure out why `gradient_checker` does not work very // well with `MatMul` and remove `TestMatMul*` in // `mnist_gradients_test` when done. GTEST_SKIP(); float A_vals[] = {1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f}; int64_t A_dims[] = {3, 3}; AbstractTensorHandlePtr A; {
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Thu Apr 13 17:32:14 GMT 2023 - 16.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_cluster_test.cc
// Check that copying it to the old remote device (named localhost) fails. TFE_TensorHandleCopyToDevice(h0_task0_new, ctx, remote_device_name, status); EXPECT_NE(TF_OK, TF_GetCode(status)) << TF_Message(status); // Copying and executing on the new remote device works. const char new_remote_device_name[] = "/job:worker/replica:0/task:1/device:CPU:0";
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Apr 14 10:03:59 GMT 2023 - 19.3K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_tensor_handle.cc
std::string& summary) const { Status status; AbstractTensorPtr resolved( // TODO(allenl): Resolve should be const, and the caches that get updated // marked mutable. const_cast<ImmediateExecutionTensorHandle*>(this)->Resolve(&status)); if (!status.ok()) { return status; } summary = resolved->SummarizeValue(); return absl::OkStatus(); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 2.1K bytes - Viewed (0) -
tensorflow/c/eager/gradients.cc
// and not for correctness. The only downside of keeping this 1 seems to be // that the gradient accumulation is unbounded and we will never // aggressively aggregate accumulated gradients to recover memory. // Revisit and fix. return 1; } // Consumes references to the tensors in the gradient_tensors list and returns // a tensor with the result. AbstractTensorHandle* TapeVSpace::AggregateGradients(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); // Pack 3 variable handles into one TFE_TensorHandle. // When remote is false, function device is placed on task0. Handle types are // REMOTE, REMOTE, LOCAL on task0. When remote is true, function device is // placed on task1, Handle types are LOCAL, REMOTE, LOCAL on task1. int num_replicas = 3; std::vector<TFE_TensorHandle*> handles = {h0, h1, h2};
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device.cc
const TFE_OpAttrs* attributes, int expected_max_outputs, TF_Status* status) { absl::optional<std::vector<MaybeParallelTensorOwned>> result; // TODO(allenl): We should remove "TPU" from these op names at the very least, // or consider other ways of packing/unpacking parallel tensors. if (operation_name == std::string("TPUReplicatedInput")) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Mar 29 22:05:31 GMT 2023 - 18.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
per_device_output_tensors.push_back(device_thread->Join(status)); // We will run every Join even if there are bad statuses in case the user // wants to recover and continue running ops on the parallel device (which // would otherwise deadlock). if (TF_GetCode(status) != TF_OK && (first_bad_status == nullptr
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1)