- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 22 for We (0.15 sec)
-
tensorflow/c/c_api.cc
status->status = session->session->Extend(std::move(graph_def)); if (!status->status.ok()) { // Contract is we always delete input_values[i]. return false; } // Note: session->session is not modified if Extend() fails, so // we only set last_num_graph_nodes if it succeeds. session->last_num_graph_nodes = num_nodes; } else { session->graph->mu.unlock();
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/array_grad.cc
absl::Span<AbstractTensorHandle*> grad_inputs) override { for (int i = 0; i < grad_outputs.size(); i++) { auto grad_input = grad_outputs[i]; // TODO(srbs): Should we add a copy contructor to AbstractTensorHandle // that takes care of this similar to `Tensor`? if (grad_input) { grad_input->Ref(); } grad_inputs[i] = grad_input; }
C++ - Registered: Tue Apr 09 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 1.6K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
gpu_options->set_allow_growth(gpu_memory_allow_growth); (*config.mutable_device_count())["CPU"] = num_cpu_devices; // TODO(b/113217601): This is needed for EagerContext::runner_ to use a // threadpool, so that we avoid the possibility of running the runner_ in the // threadpool of GPU event mgr, as that can trigger more callbacks to be // scheduled on that same threadpool, causing a deadlock in cases where the
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/posix/posix_filesystem.cc
TF_SetStatusFromIOError(status, errno, dst); return; } } else if (S_ISDIR(st.st_mode)) { TF_SetStatus(status, TF_FAILED_PRECONDITION, "target path is a directory"); return; } // We cannot rename directories yet, so prevent this. if (stat(src, &st) != 0) { TF_SetStatusFromIOError(status, errno, src); return; } else if (S_ISDIR(st.st_mode)) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Sun Mar 24 20:08:23 GMT 2024 - 15.8K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// Wait until all pending nodes have completed since they may have a // reference to default_cancellation_manager_. We ignore the status return // since we already have a bad status to propagate. TFE_ContextAsyncWait(context, await_status.get()); // Reset the cancellation manager on a bad status. Otherwise we'll cancel // all future operations. default_cancellation_manager_ = absl::make_unique<CancellationManager>(); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/gradient_checker.cc
vector<AbstractTensorHandle*> theta_inputs(inputs.size()); for (int i{}; i < inputs.size(); ++i) { theta_inputs[i] = inputs[i]; } AbstractTensorHandle* theta = theta_inputs[input_index]; // parameter we are grad checking // Convert from AbstractTensor to TF_Tensor. TF_Tensor* theta_tensor; TF_RETURN_IF_ERROR(GetValue(theta, &theta_tensor)); // Get number of elements and fill data.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 7.3K bytes - Viewed (0) -
tensorflow/c/c_api_function.cc
if (iter == input_nodes.end()) { // This node is not referenced in inputs. Add it to the body. body_nodes->push_back(node); } else { // This node is referenced in inputs. Currently, we place an // artificial restriction and require that when num_opers=-1, such // nodes must have a single output. if (node->num_outputs() != 1) { return InvalidArgument(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 13.6K bytes - Viewed (2) -
tensorflow/c/eager/c_api_unified_experimental_test.cc
TF_DeleteOutputList(func_outputs); } /** * We traced so far this function: * * def two_adds(a, b): * my_add1 = a + b * my_add2 = b + b * return my_add1, my_add2 * * Now we will execute this function with an eager context: * * output1, output2 = two_adds(2.0, 3.0) * * and check that we got 5.0 and 6.0 as results. */ // Build eager context.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri May 19 21:44:52 GMT 2023 - 39.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api_distributed_test.cc
TFE_ContextSetServerDef(ctx, 0, serialized.data(), serialized.size(), status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); // Use large matrices so that RPCs don't return before we get a chance // to call TFE_DeleteContext. TFE_TensorHandle* h0_task0 = TestMatrixTensorHandle100x100(ctx); TFE_TensorHandle* h1_task0 = TestMatrixTensorHandle100x100(ctx); const char remote_device_name[] =
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/eager/c_api_test.cc
if (forward_input) { // Since the input was forwarded, we released the input handle right away // and hence expect the input to be forwarded to the return tensor. EXPECT_EQ(orig_ptr, TF_TensorData(t)); } else { // In async mode we expect forwarding to work without releasing the input // handle since by the time the kernel is executed we have released the // handle in the client code.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 03 20:50:20 GMT 2023 - 94.6K bytes - Viewed (1)