- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 37 for Leach (0.22 sec)
-
ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc
# installed into a virtual environment, and then that venv is used to run all # bazel tests with a special flag "--define=no_tensorflow_py_deps=true", which # drops all the bazel dependencies for each py_test; this makes all the tests # use the wheel's TensorFlow installation instead of the one made available # through bazel. This must be done in a different root directory, //bazel_pip/...,
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Nov 21 12:25:39 GMT 2023 - 5.8K bytes - Viewed (0) -
tensorflow/c/c_api_internal.h
// Maps from name of an operation to the Node* in 'graph'. std::unordered_map<tensorflow::string, tensorflow::Node*> name_map TF_GUARDED_BY(mu); // The keys of this map are all the active sessions using this graph. Each // value records whether the graph has been mutated since the corresponding // session has been run (this is detected in RecordMutation function). If the // string is empty, no mutation has occurred. Otherwise the string is a
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Sat May 13 00:49:12 GMT 2023 - 7.6K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc
break; case FetchState::FINISHED: return TF_SetStatus(status, TF_OK, ""); } } return TF_SetStatus( status, TF_INTERNAL, "Control flow should never reach the end of RamFileBlockCache::Fetch."); } int64_t RamFileBlockCache::Read(const std::string& filename, size_t offset, size_t n, char* buffer, TF_Status* status) { if (n == 0) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 16 01:39:09 GMT 2020 - 11.1K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
// The cache has space for `block_count` blocks. The loop with i = 0 should // fill the cache, and the loop with i = 1 should be all cache hits. The // fetcher checks that it is called once and only once for each offset (to // fetch the corresponding block). for (int i = 0; i < 2; i++) { for (int j = 0; j < block_count; j++) { TF_EXPECT_OK(ReadCache(&cache, "", block_size * j, block_size, &out)); } }
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:57 GMT 2021 - 23.2K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/modular_filesystem_test.cc
namespace tensorflow { namespace { using ::tensorflow::error::Code; // As we need to test multiple URI schemes we need a parameterized test. // Furthermore, since each test creates and deletes files, we will use the same // fixture to create new directories in `SetUp`. Each directory will reside in // `::testing::TempDir()`, will use a RNG component and the test name. This // ensures that two consecutive runs are unlikely to clash.
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri May 27 20:25:58 GMT 2022 - 71K bytes - Viewed (0) -
tensorflow/c/eager/c_api_unified_experimental_test.cc
TF_Tensor* f_t = TFE_TensorHandleResolve(handle, s); ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s); memcpy(&result_data[0], TF_TensorData(f_t), TF_TensorByteSize(f_t)); // Verify results for each output for (int j = 0; j < 4; j++) { ASSERT_EQ(result_data[j], expected_outputs[idx][j]); } TF_DeleteTensor(f_t); } // Free memory associated with add and MatMul outputs
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri May 19 21:44:52 GMT 2023 - 39.1K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
EXPECT_EQ(X, dxy_dyValue); TF_DeleteTensor(fetchValues[0]); TF_DeleteTensor(fetchValues[1]); } // REGISTER_OP for CApiAttributesTest test cases. // Registers two ops, each with a single attribute called 'v'. // The attribute in one op will have a type 'type', the other // will have list(type). #define ATTR_TEST_REGISTER_OP(type) \
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 96.9K bytes - Viewed (3) -
tensorflow/c/c_api_experimental.cc
// (as in the case of ConstOp kernel creation on GPU, which involves copying a // CPU tensor to GPU). // Setting a larger thread pool does not help with the Swift caller, as we use // a different TFE context for each thread of execution (for running graph // functions, and their send/recvs corountines). config.set_inter_op_parallelism_threads(1); TF_Buffer* ret = TF_NewBuffer(); TF_CHECK_OK(MessageToBuffer(config, ret));
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// // First we filter the tape to just the subset of operations we want to // differentiate. In the process of doing so we count how many times each Tensor // is used as an input to an op (so we know when we're done computing gradients // for that Tensor). We also count, for each tape entry, how many of its output // Tensors need gradients to be computed (Tensors which are not used do not need // any gradients to be computed). //
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
configure.py
'device at: https://developer.nvidia.com/cuda-gpus. Each capability ' 'can be specified as "x.y" or "compute_xy" to include both virtual and' ' binary GPU code, or as "sm_xy" to only include the binary ' 'code.\nPlease note that each additional compute capability ' 'significantly increases your build time and binary size, and that '
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 18:25:36 GMT 2024 - 53.8K bytes - Viewed (0)