- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 28 for Falls (0.16 sec)
-
tensorflow/c/eager/tfe_op_internal.h
// Wraps a pointer to an operation implementation. // // WARNING: Since the underlying object could be ref-counted a user of this // interface cannot destruct the underlying operation object. Instead, call // TFE_DeleteOp who calls Release() on the operation pointer and deletes // the TFE_Op structure. typedef struct TFE_Op TFE_Op; namespace tensorflow { DEFINE_CONVERSION_FUNCTIONS(tensorflow::ImmediateExecutionOperation, TFE_Op);
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Jun 17 19:39:13 GMT 2020 - 1.4K bytes - Viewed (0) -
tensorflow/c/eager/tfe_tensorhandle_internal.h
// // WARNING: Since the underlying object could be ref-counted a user of this // interface cannot destruct the underlying handle object. Instead, call // TFE_DeleteTensorHandle who calls Release() on the handle pointer and deletes // the TFE_TensorHandle structure. typedef struct TFE_TensorHandle TFE_TensorHandle; namespace tensorflow {
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Jun 17 19:39:13 GMT 2020 - 1.6K bytes - Viewed (0) -
WORKSPACE
) load("@pypi//:requirements.bzl", "install_deps") install_deps() # Initialize the TensorFlow repository and all dependencies. # # The cascade of load() statements and tf_workspace?() calls works around the # restriction that load() statements need to be at the top of .bzl files. # E.g. we can not retrieve a new repository with http_archive and then load() # a macro from that repository in the same file.
Plain Text - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Apr 05 22:27:48 GMT 2024 - 3K bytes - Viewed (2) -
tensorflow/c/eager/gradients.cc
// a tensor with the result. AbstractTensorHandle* AggregateGradients( gtl::ArraySlice<AbstractTensorHandle*> gradient_tensors) const override; // Calls the passed-in backward function. // op_type is the op's name provided in RecordOperation. Status CallBackwardFunction( const string& op_type, GradientFunction* gradient_function,
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0) -
tensorflow/c/experimental/grappler/grappler.cc
// This file extends/implements core graph optimizer base classes in terms of // the C API defined in grappler.h. A class "CSomething" represents a // "Something" that can be manipulated via calls in the C interface and a C // struct called "TP_Something". #include "tensorflow/c/experimental/grappler/grappler.h" #include <algorithm> #include <cstddef> #include <cstring> #include <string>
C++ - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Wed Sep 06 19:12:29 GMT 2023 - 15K bytes - Viewed (1) -
tensorflow/c/eager/c_api_experimental.h
// ops return an error. This call clears the error state and re-enables // execution of newly issued ops. // // Note that outputs of discarded ops remain in a corrupt state and should not // be used for future calls. // TODO(agarwal): mark the affected handles and raise errors if they are used. TF_CAPI_EXPORT extern void TFE_ExecutorClearError(TFE_Executor*); // Sets a custom Executor for the current thread. All nodes created by this
C - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/repack_libtensorflow.sh
# under the root of srcjars jars created by bazel, rather than under # the maven-style src/main/java subdirectory. # # Bazel manages annotation generated source as follows: First, it # calls javac with options that create generated files under a # bazel-out directory. Next, it archives the generated source files # into a srcjar directly under the root. There doesn't appear to be a
Shell Script - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Mon Sep 18 14:52:45 GMT 2023 - 2.6K bytes - Viewed (0) -
RELEASE.md
`compression=None` to `tf.data.experimental.service.distribute(...)`. * `tf.data.Dataset.batch()` now supports `num_parallel_calls` and `deterministic` arguments. `num_parallel_calls` is used to indicate that multiple input batches should be computed in parallel. With `num_parallel_calls` set, `deterministic` is used to indicate that outputs can be obtained in the non-deterministic order.
Plain Text - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Wed Apr 03 20:27:38 GMT 2024 - 727.4K bytes - Viewed (8) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
std::list<size_t> calls; auto fetcher = [&calls, block_size](const string& filename, size_t offset, size_t n, char* buffer, TF_Status* status) -> int64_t { EXPECT_EQ(n, block_size); EXPECT_FALSE(calls.empty()) << "at offset = " << offset; if (!calls.empty()) { EXPECT_EQ(offset, calls.front()); calls.pop_front(); }
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:57 GMT 2021 - 23.2K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/expiring_lru_cache_test.cc
TEST(ExpiringLRUCacheTest, LookupOrCompute) { // max_age of 0 means we should always compute. uint64 num_compute_calls = 0; tf_gcs_filesystem::ExpiringLRUCache<int>::ComputeFunc compute_func = [&num_compute_calls](const string& key, int* value, TF_Status* status) { *value = num_compute_calls; num_compute_calls++; return TF_SetStatus(status, TF_OK, ""); };
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 09 19:31:22 GMT 2020 - 7.1K bytes - Viewed (0)