- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 25 for have (0.18 sec)
-
CONTRIBUTING.md
***NOTE***: Only original source code from you and other people that have signed the CLA can be accepted into the main repository. ### Contributing code If you have improvements to TensorFlow, send us your pull requests! For those just getting started, GitHub has a [how-to](https://help.github.com/articles/using-pull-requests/).
Plain Text - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Mar 21 11:45:51 GMT 2024 - 15.6K bytes - Viewed (0) -
tensorflow/c/experimental/grappler/grappler.h
// * Refer to "TensorFlow Versioning Strategy" section at // https://github.com/tensorflow/community/pull/257/files. // * Note that the API is still under active development and doesn't have // versioning guarantees yet. // * `void* ext` is a free-form field that can be populated by // a plugin in `TP_*` structs or potential future extension points . // // Example usage: //
C - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Wed Aug 03 18:08:43 GMT 2022 - 12.5K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
// Causes the calling thread to block till all ops dispatched in this executor // have been executed. Note that "execution" here refers to kernel execution / // scheduling of copies, etc. Similar to sync execution, it doesn't guarantee // that lower level device queues (like GPU streams) have been flushed. // // This call may not block for execution of ops enqueued concurrently with this // call.
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
int64_t expected_stride = 1; for (int i = ndim - 1; i >= 0; --i) { // Empty tensors are always compact regardless of strides. if (shape_arr[i] == 0) return true; // Note that dimensions with size=1 can have any stride. if (shape_arr[i] != 1 && stride_arr[i] != expected_stride) { valid = false; } expected_stride *= shape_arr[i]; } return valid; } } // namespace
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/eager/c_api.h
// things like the available devices, resource manager etc. // TFE_Context must outlive all tensor handles created using it. In other // words, TFE_DeleteContext() must be called after all tensor handles have // been deleted (with TFE_DeleteTensorHandle). // // TODO(ashankar): Merge with TF_Session? typedef struct TFE_Context TFE_Context; TF_CAPI_EXPORT extern TFE_Context* TFE_NewContext(
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 22.8K bytes - Viewed (1) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
TEST(RamFileBlockCacheTest, ParallelReads) { // This fetcher won't respond until either `callers` threads are calling it // concurrently (at which point it will respond with success to all callers), // or 10 seconds have elapsed (at which point it will respond with an error). const int callers = 4; BlockingCounter counter(callers); auto fetcher = [&counter](const string& filename, size_t offset, size_t n,
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:57 GMT 2021 - 23.2K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc
block->lru_iterator = lru_list_.begin(); } // Check for inconsistent state. If there is a block later in the same file // in the cache, and our current block is not block size, this likely means // we have inconsistent state within the cache. Note: it's possible some // incomplete reads may still go undetected. if (block->data.size() < block_size_) { Key fmax = std::make_pair(key.first, std::numeric_limits<size_t>::max());
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 16 01:39:09 GMT 2020 - 11.1K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.h
// Blocks until the previous `StartExecute` has run `TFE_Execute` on each // device. If is_async=false (constructor argument) this means the ops have // run and have results. If is_async=true it means that all of the // device-specific executors have scheduled the op. // // Accepts inferred shapes for outputs (`expected_output_shapes`), which if // fully defined will avoid querying the shapes of the underlying
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Tue Apr 25 15:21:13 GMT 2023 - 12.9K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
std::unique_ptr<TF_Status, decltype(&TF_DeleteStatus)> await_status( TF_NewStatus(), TF_DeleteStatus); // Wait until all pending nodes have completed since they may have a // reference to default_cancellation_manager_. We ignore the status return // since we already have a bad status to propagate. TFE_ContextAsyncWait(context, await_status.get()); // Reset the cancellation manager on a bad status. Otherwise we'll cancel
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/experimental/filesystem/modular_filesystem.cc
#include "tensorflow/core/platform/file_system_helper.h" #include "tsl/platform/errors.h" // TODO(b/139060984): After all filesystems are converted, all calls to // methods from `FileSystem` will have to be replaced to calls to private // methods here, as part of making this class a singleton and the only way to // register/use filesystems. namespace tensorflow { using UniquePtrTo_TF_Status =
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Sep 06 19:12:29 GMT 2023 - 23.1K bytes - Viewed (0)