- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for So (0.15 sec)
-
tensorflow/c/c_api.cc
shape_handle, static_cast<DataType>(types[i])); } ic->set_output_handle_shapes_and_types(output.index, shape_and_type_vec); } // Helpers for loading a TensorFlow plugin (a .so file). Status LoadDynamicLibrary(const char* library_filename, void** result, const void** buf, size_t* len); // TODO(josh11b,mrry): Change Session to be able to use a Graph*
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 102.3K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/posix/posix_filesystem.cc
static_cast<off_t>(offset))}; if (r > 0) { dst += r; offset += static_cast<uint64_t>(r); n -= r; // safe as 0 < r <= n so n will never underflow read += r; } else if (r == 0) { TF_SetStatus(status, TF_OUT_OF_RANGE, "Read fewer bytes than requested"); break; } else if (errno == EINTR || errno == EAGAIN) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Sun Mar 24 20:08:23 GMT 2024 - 15.8K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
gpu_options->set_allow_growth(gpu_memory_allow_growth); (*config.mutable_device_count())["CPU"] = num_cpu_devices; // TODO(b/113217601): This is needed for EagerContext::runner_ to use a // threadpool, so that we avoid the possibility of running the runner_ in the // threadpool of GPU event mgr, as that can trigger more callbacks to be // scheduled on that same threadpool, causing a deadlock in cases where the
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/array_grad_test.cc
TF_RETURN_IF_ERROR( ops::IdentityN(ctx, inputs, absl::MakeSpan(temp_outputs), "IdentityN")); // Although, `ops::IdentityN` returns 2 tensors, the first tensor isn't needed // for computing gradient so we could safely drop it. outputs[0] = temp_outputs[1]; temp_outputs[0]->Unref(); return absl::OkStatus(); } class CppGradients : public ::testing::TestWithParam<std::tuple<const char*, bool, bool>> {
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5K bytes - Viewed (0) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
// `gradient_checker` only works with model that returns only 1 tensor. // Although, `ops::SparseSoftmaxCrossEntropyWithLogits` returns 2 tensors, the // second tensor isn't needed for computing gradient so we could safely drop // it. outputs[0] = loss; backprop->Unref(); return absl::OkStatus(); } Status BiasAddModel(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs,
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 8.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
CancellationManager* cancellation_manager_ TF_GUARDED_BY(execution_mutex_); // Outputs: std::vector<TensorHandlePtr> op_outputs_ TF_GUARDED_BY(execution_mutex_); // TF_Status is an incomplete type and so can't be stack allocated. To avoid // unnecessary allocations each Execute call, we keep one heap-allocated // version for the thread. StatusPtr status_ TF_GUARDED_BY(execution_mutex_); const std::string device_;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
worker_server1.release(); worker_server2.release(); // Disable the test graph pass so it does not affect other test cases. GraphErrorInjectionPass::enabled_ = false; } string VariableAddFunctionWithGraphError() { string signature = VariableAddFunctionSignature(); // Replace the node 'read0' with 'read0_maybe_with_graph_error', so that the // error injecting pass can identify and introduce graph pass errors.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/c_api_function.cc
"' has a reference type ", DataTypeString(dt)) : absl::OkStatus(); } // Converts `ninputs` and `inputs` into `inputs_tensors` and `input_nodes` and // does various checks while doing so. `input_nodes` will contain the same // information as input_tensors just in a different structure to make // following processing easier. TODO(iga): Simplify this nested structure. Status ProcessInputs(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 13.6K bytes - Viewed (2) -
tensorflow/c/eager/gradient_checker.cc
if (num_dims_out == 0) { outputs[0] = model_out.release(); return absl::OkStatus(); } // Else, reduce sum the output to get a scalar // Will sum all dimensions, so get a Tensor containing [0,...,num_dims_out-1]. AbstractTensorHandlePtr sum_dims; { vector<int32_t> vals(num_dims_out); int64_t vals_shape[] = {num_dims_out}; Range(&vals, 0, num_dims_out);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 7.3K bytes - Viewed (0) -
tensorflow/c/eager/unified_api_testutil.cc
if (use_function) { const char* fn_name = "test_fn"; core::RefCountPtr<AbstractFunction> scoped_func; // Returning null tensors from a tf.function is not supported, so we keep // track of indices in the model's outputs are nullptr in this set. // The FunctionDef only outputs the non-null tensors. We later pad the // function op outputs to have nullptrs at the `null_indices`.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Feb 27 13:57:45 GMT 2024 - 5.7K bytes - Viewed (0)