- Sort Score
- Result 10 results
- Languages All
Results 1 - 3 of 3 for queue (0.15 sec)
-
tensorflow/c/c_api_experimental.h
// `tensor_id`, associated with `session`. There must be a graph node named // "fifo_queue_enqueue_<tensor_id>", to be executed by this API call. It reads // from a placeholder node "arg_tensor_enqueue_<tensor_id>". // // `tensor` is still owned by the caller. This call will be blocked if the queue // has reached its capacity, and will be unblocked when the queued tensors again // drop below the capacity due to dequeuing. //
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 15.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
// have been executed. Note that "execution" here refers to kernel execution / // scheduling of copies, etc. Similar to sync execution, it doesn't guarantee // that lower level device queues (like GPU streams) have been flushed. // // This call may not block for execution of ops enqueued concurrently with this // call. TF_CAPI_EXPORT extern void TFE_ExecutorWaitForAllPendingNodes(
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.h
// re-use a thread. std::vector<std::unique_ptr<DeviceThread>> device_threads_; // A cancellation manager to use if the caller does not provide one. When ops // are executed asynchronously this must outlive the queued op, so it can't be // function-local to Execute. mutable std::unique_ptr<CancellationManager> default_cancellation_manager_; }; // Contains a tuple of tensors, one on each of the `underlying_devices_` of the
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 25 15:21:13 GMT 2023 - 12.9K bytes - Viewed (0)