- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 14 for sync (0.25 sec)
-
tensorflow/c/experimental/filesystem/modular_filesystem.cc
ops_->flush(file_.get(), plugin_status.get()); return StatusFromTF_Status(plugin_status.get()); } Status ModularWritableFile::Sync() { if (ops_->sync == nullptr) return Flush(); UniquePtrTo_TF_Status plugin_status(TF_NewStatus(), TF_DeleteStatus); ops_->sync(file_.get(), plugin_status.get()); return StatusFromTF_Status(plugin_status.get()); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Sep 06 19:12:29 GMT 2023 - 23.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api.h
TFE_DEVICE_PLACEMENT_SILENT_FOR_INT32 = 3, } TFE_ContextDevicePlacementPolicy; // LINT.ThenChange(//tensorflow/c/eager/immediate_execution_context.h) // Sets the default execution mode (sync/async). Note that this can be // overridden per thread using TFE_ContextSetExecutorForThread. TF_CAPI_EXPORT extern void TFE_ContextOptionsSetAsync(TFE_ContextOptions*,
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 22.8K bytes - Viewed (1) -
tensorflow/c/eager/c_api_experimental_test.cc
ASSERT_TRUE(TF_GetCode(status) == TF_OK) << TF_Message(status); TF_DeleteFunction(fn); for (bool async : {false, true, false}) { TFE_Executor* old_executor = TFE_ContextGetExecutorForThread(ctx); TFE_Executor* executor = TFE_NewExecutor( /*is_async=*/async, /*enable_streaming_enqueue=*/true, /*in_flight_nodes_limit=*/0); TFE_ContextSetExecutorForThread(ctx, executor);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 03 03:14:26 GMT 2023 - 31.5K bytes - Viewed (1) -
tensorflow/c/experimental/filesystem/plugins/gcs/gcs_filesystem.cc
gcs_file->object.c_str()); if (TF_GetCode(status) != TF_OK) return; gcs_file->sync_need = false; } else { TF_SetStatus(status, TF_OK, ""); } } void Sync(const TF_WritableFile* file, TF_Status* status) { auto gcs_file = static_cast<GCSFile*>(file->plugin_file); TF_VLog(3, "Sync: gs://%s/%s", gcs_file->bucket.c_str(), gcs_file->object.c_str()); Flush(file, status); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Aug 23 06:55:53 GMT 2023 - 46.9K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
// executed in sequence. Assigning nodes to different executors allows executing // nodes in parallel. // in_flight_nodes_limit: when is_async is true, this value controls the // maximum number of in flight async nodes. Enqueuing of additional async ops // after the limit is reached blocks until some inflight nodes finishes. // The effect is bounding the memory held by inflight TensorHandles that are
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/posix/posix_filesystem.cc
if (fflush(posix_file->handle) != 0) TF_SetStatusFromIOError(status, errno, posix_file->filename); } static void Sync(const TF_WritableFile* file, TF_Status* status) { // For historical reasons, this does the same as `Flush` at the moment. // TODO(b/144055243): This should use `fsync`/`sync`. Flush(file, status); } static void Close(const TF_WritableFile* file, TF_Status* status) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Sun Mar 24 20:08:23 GMT 2024 - 15.8K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_context.h
#include "tensorflow/core/util/device_name_utils.h" namespace tensorflow { class EagerExecutor; class EagerContext; class CustomDevice; class CustomDeviceOpHandler; class Device; // LINT.IfChange // Note: Keep in sync with exported copy of enum in eager/c_api.h. enum ContextDevicePlacementPolicy { // Running operations with input tensors on the wrong device will fail. DEVICE_PLACEMENT_EXPLICIT = 0,
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 06 08:34:00 GMT 2023 - 12.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api.cc
} // Note: this function looks up a thread local policy. So it should be called in // the appropriate client thread. In particular, in async mode, it may not be // safe to call this function from the async EagerExecutor threads. extern TFE_ContextDevicePlacementPolicy TFE_ContextGetDevicePlacementPolicy( TFE_Context* ctx) { return static_cast<TFE_ContextDevicePlacementPolicy>(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 12 20:00:09 GMT 2024 - 43.9K bytes - Viewed (2) -
tensorflow/c/eager/c_api_experimental.cc
tensorflow::unwrap(cancellation_manager)); status->status = absl::OkStatus(); } TFE_Executor* TFE_NewExecutor(bool is_async, bool enable_streaming_enqueue, int in_flight_nodes_limit) { return new TFE_Executor(is_async, enable_streaming_enqueue, in_flight_nodes_limit); } void TFE_DeleteExecutor(TFE_Executor* executor) { delete executor; }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 11 23:52:39 GMT 2024 - 35.9K bytes - Viewed (3) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
const int in_flight_nodes_limit) : status_(TF_NewStatus()), // If the context's default exector is set to async, re-using that in // each thread would cause collectives to deadlock. For consistency we // create a new sync executor for every thread. // // TODO(allenl): We should have an async API that works with the
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1)