- Sort Score
- Result 10 results
- Languages All
Results 1 - 5 of 5 for re (0.2 sec)
-
tensorflow/c/experimental/filesystem/plugins/gcs/expiring_lru_cache_test.cc
EXPECT_EQ(num_compute_calls, 5); cache2.LookupOrCompute("d", &value, compute_func, &status); TF_EXPECT_OK(status.status); EXPECT_EQ(value, 3); EXPECT_EQ(num_compute_calls, 5); // Re-read "a", ensure it is re-computed. cache2.LookupOrCompute("a", &value, compute_func, &status); TF_EXPECT_OK(status.status); EXPECT_EQ(value, 5); EXPECT_EQ(num_compute_calls, 6); } TEST(ExpiringLRUCacheTest, Clear) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 09 19:31:22 GMT 2020 - 7.1K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc
} auto begin = data.begin(); if (offset > pos) { // The block begins before the slice we're reading. begin += offset - pos; } auto end = data.end(); if (pos + data.size() > offset + n) { // The block extends past the end of the slice we're reading. end -= (pos + data.size()) - (offset + n); } if (begin < end) { size_t bytes_to_copy = end - begin;
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 16 01:39:09 GMT 2020 - 11.1K bytes - Viewed (0) -
tensorflow/c/eager/c_api_test.cc
} else { // In sync mode, forwarding can't really happen since the client code will // have a reference count on the input tensor while the kernel is being // executed and thus it cannot be re-used for the return tensor. EXPECT_NE(orig_ptr, TF_TensorData(t)); } } ASSERT_EQ(TF_OK, TF_GetCode(status)) << TF_Message(status); TFE_DeleteTensorHandle(m); TFE_DeleteTensorHandle(retval);
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Aug 03 20:50:20 GMT 2023 - 94.6K bytes - Viewed (1) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
explicit DeviceThread(const std::string& device, const bool is_async, const int in_flight_nodes_limit) : status_(TF_NewStatus()), // If the context's default exector is set to async, re-using that in // each thread would cause collectives to deadlock. For consistency we // create a new sync executor for every thread. //
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_experimental.cc
tensorflow::ImmediateExecutionTensorHandle* unwrapped_handle = tensorflow::unwrap(handles[i]); if (tensorflow::CustomDeviceTensorHandle::classof(unwrapped_handle)) { // One of the inputs we're trying to pack is on a custom device. We'll let // the first custom device we see handle all of the packing. auto* custom_device_handle = tensorflow::down_cast<tensorflow::CustomDeviceTensorHandle*>(
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Apr 11 23:52:39 GMT 2024 - 35.9K bytes - Viewed (3)