- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for Fuller (0.19 sec)
-
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
std::vector<std::unique_ptr<Thread>> threads; threads.reserve(callers); for (int i = 0; i < callers; i++) { threads.emplace_back( Env::Default()->StartThread({}, "caller", [block_size, &cache, i]() { std::vector<char> out; TF_EXPECT_OK( ReadCache(&cache, "a", i * block_size, block_size, &out)); std::vector<char> x(block_size, 'x');
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:57 GMT 2021 - 23.2K bytes - Viewed (0) -
tensorflow/c/eager/c_api.h
typedef struct TFE_TensorHandle TFE_TensorHandle; TF_CAPI_EXPORT extern TFE_TensorHandle* TFE_NewTensorHandle(const TF_Tensor* t, TF_Status* status); // Indicates that the caller will not be using `h` any more. TF_CAPI_EXPORT extern void TFE_DeleteTensorHandle(TFE_TensorHandle* h); TF_CAPI_EXPORT extern TF_DataType TFE_TensorHandleDataType(TFE_TensorHandle* h);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 22.8K bytes - Viewed (1) -
tensorflow/c/eager/parallel_device/parallel_device_lib.h
// before `StartExecute` is called again. Using `StartExecute` with `Join` // allows the caller to schedule computation on multiple ParallelDevices // without sequencing those operations (first call `StartExecute` on each // parallel device, then call `Join` on each; even if some of the `Join`s // return a bad status the caller must run all of the `Join`s or any future // `StartExecute`s will deadlock). //
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 25 15:21:13 GMT 2023 - 12.9K bytes - Viewed (0) -
tensorflow/c/experimental/next_pluggable_device/c_api.cc
TF_Buffer* ProcessGetKeyValueResult(absl::StatusOr<std::string> value, TF_Status* status) { status->status = value.status(); if (!value.ok()) { return nullptr; } // Caller is responsible to call `TF_DeleteBuffer` to release the buffer. TF_Buffer* result = TF_NewBuffer(); const std::string& value_str = *value; void* data = malloc(value_str.length());
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Tue Jan 09 00:52:04 GMT 2024 - 13.9K bytes - Viewed (1) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
// finished, if any. // // `cancellation_manager` must live until after `Join` finishes and pending // `is_async` operations finish. In addition to allowing the caller to cancel // the operation, its `StartCancel` method will be called if op execution // fails on any device in order to cancel the others. void StartExecute(TFE_Context* context, const char* operation_name,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/c_api_experimental.h
// The returned string is heap-allocated, and caller should call free() on it. TF_CAPI_EXPORT extern const char* TF_GraphDebugString(TF_Graph* graph, size_t* len); // Returns the function content in a human-readable format, with length set in // `len`. The format is subject to change in the future. // The returned string is heap-allocated, and caller should call free() on it. //
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 15.1K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// scheduled on that same threadpool, causing a deadlock in cases where the // caller of event_mgr->ThenExecute() blocks on the completion of the callback // (as in the case of ConstOp kernel creation on GPU, which involves copying a // CPU tensor to GPU). // Setting a larger thread pool does not help with the Swift caller, as we use // a different TFE context for each thread of execution (for running graph
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// Fetches the current Jacobian-vector product associated with `tensor_id`, or // a nullptr if none is available. // // Returns a borrowed reference, i.e. does not run VSpace::MarkAsResult on its // return value. The caller should increment the reference count before // deleting the ForwardAccumulator or calling DeleteGradient if keeping a // persistent reference to a non-null result. Gradient* FetchJVP(int64_t tensor_id);
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
tensorflow/c/eager/c_api_experimental.h
TFE_MonitoringCounterCell* cell); // APIs for Counter without label. typedef struct TFE_MonitoringCounter0 TFE_MonitoringCounter0; // Returns a new Counter metric object. The caller should manage lifetime of // the object. Using duplicate metric name will crash the program with fatal // error. TF_CAPI_EXPORT extern TFE_MonitoringCounter0* TFE_MonitoringNewCounter0(
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_context.h
// Configure graph collection in RunMetadata. virtual void SetShouldStoreGraphs(bool value) = 0; // Return the collected RunMetadata. This method will transfer the ownership // to the caller. virtual std::unique_ptr<RunMetadata> ExportRunMetadata() = 0; // For LLVM style RTTI. static bool classof(const AbstractContext* ptr) { return ptr->getKind() == kEager || ptr->getKind() == kTfrt; }
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 06 08:34:00 GMT 2023 - 12.3K bytes - Viewed (0)