- Sort Score
- Result 10 results
- Languages All
Results 1 - 7 of 7 for Sync (0.15 sec)
-
tensorflow/c/experimental/filesystem/plugins/posix/posix_filesystem.cc
if (fflush(posix_file->handle) != 0) TF_SetStatusFromIOError(status, errno, posix_file->filename); } static void Sync(const TF_WritableFile* file, TF_Status* status) { // For historical reasons, this does the same as `Flush` at the moment. // TODO(b/144055243): This should use `fsync`/`sync`. Flush(file, status); } static void Close(const TF_WritableFile* file, TF_Status* status) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Sun Mar 24 20:08:23 GMT 2024 - 15.8K bytes - Viewed (0) -
ci/official/upload.sh
# contents. About arguments to gsutil: # "gsutil -m rsync" runs in parallel. # "gsutil rsync -r" is recursive and makes directories work. # "gsutil rsync -d" is "sync and delete files from destination if not present in source" DOWNLOADS="$(mktemp -d)" mkdir -p "$DOWNLOADS" gsutil -m rsync -r "$TFCI_ARTIFACT_STAGING_GCS_URI" "$DOWNLOADS" ls "$DOWNLOADS"
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Jan 24 20:52:12 GMT 2024 - 2.8K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
// Causes the calling thread to block till all ops dispatched in this executor // have been executed. Note that "execution" here refers to kernel execution / // scheduling of copies, etc. Similar to sync execution, it doesn't guarantee // that lower level device queues (like GPU streams) have been flushed. // // This call may not block for execution of ops enqueued concurrently with this // call.
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/c_api.cc
handle->TypeString(), " tensor handle."); return nullptr; } tensorflow::Device* device(handle->device()); if (device != nullptr) { status->status = device->Sync(); if (!status->status.ok()) { return nullptr; } } const tensorflow::Tensor* tensor; status->status = handle->Tensor(&tensor); if (!status->status.ok()) { return nullptr; }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 12 20:00:09 GMT 2024 - 43.9K bytes - Viewed (2) -
RELEASE.md
* `tf.distribute`: * Support added for global sync `BatchNormalization` by using the newly added `tf.keras.layers.experimental.SyncBatchNormalization` layer. This layer will sync `BatchNormalization` statistics every step across all replicas taking part in sync training. * Performance improvements for GPU multi-worker distributed training using
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Mon Apr 29 19:17:57 GMT 2024 - 727.7K bytes - Viewed (8) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
: status_(TF_NewStatus()), // If the context's default exector is set to async, re-using that in // each thread would cause collectives to deadlock. For consistency we // create a new sync executor for every thread. // // TODO(allenl): We should have an async API that works with the // parallel device. device_(device), executor_(
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
TFE_TensorHandle* h0 = TestVariable(ctx, 1.0, task1_name); TFE_TensorHandle* h1 = TestVariable(ctx, 2.0, task2_name); TFE_TensorHandle* h2 = TestVariable(ctx, 3.0, task0_name); // Add a sync point to make sure that variables have been initialized // before the function execution starts. TFE_ContextAsyncWait(ctx, status); EXPECT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0)