- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for can (0.13 sec)
-
tensorflow/c/eager/c_api_cluster_test.cc
TEST(CAPI, TestRemoteExecuteUpdateServerDefResourceAccessAsync) { TestRemoteExecuteUpdateServerDefResourceAccess(true); } void TestRemoteExecuteUpdateServerDefWithFailures(bool async) { // Fail fast on GetStatus requests so we can get errors instead of timeout // when updating cluster with non-exsitent worker tensorflow::setenv("GRPC_FAIL_FAST", "TRUE", /*overwrite=*/1); tensorflow::ServerDef server_def = GetServerDef(2);
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Apr 14 10:03:59 GMT 2023 - 19.3K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// TODO(b/113217601): This is needed for EagerContext::runner_ to use a // threadpool, so that we avoid the possibility of running the runner_ in the // threadpool of GPU event mgr, as that can trigger more callbacks to be // scheduled on that same threadpool, causing a deadlock in cases where the // caller of event_mgr->ThenExecute() blocks on the completion of the callback
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 29.4K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib_test.cc
ParallelTensor::FromTensorHandles(parallel_device, std::move(mixed_handles), status.get()); ASSERT_TRUE(TF_GetCode(status.get()) == TF_OK); // Can't take the shape of a parallel tensor with varying numbers of axes, but // running operations on them is OK. TF_ASSERT_OK(unknown_length_vector->Shape(&shape)); EXPECT_THAT(*shape, ElementsAre(-1));
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Jul 08 23:47:35 GMT 2021 - 15.3K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
int64_t expected_stride = 1; for (int i = ndim - 1; i >= 0; --i) { // Empty tensors are always compact regardless of strides. if (shape_arr[i] == 0) return true; // Note that dimensions with size=1 can have any stride. if (shape_arr[i] != 1 && stride_arr[i] != expected_stride) { valid = false; } expected_stride *= shape_arr[i]; } return valid; } } // namespace
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/eager/custom_device_test.cc
num_retvals = 1; executed = false; TFE_Execute(matmul.get(), &retval, &num_retvals, status.get()); EXPECT_FALSE(executed); ASSERT_FALSE(TF_GetCode(status.get()) == TF_OK); // Custom devices can refuse to do type-based dispatch (as hcustom1 is // configured to do) matmul.reset(MatMulOp(context.get(), hcustom1.get(), hcpu.get())); num_retvals = 1; executed = false;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Aug 27 23:39:24 GMT 2020 - 18.4K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.cc
// Copy the relevant portion of the block into the result buffer. const auto& data = block->data; if (offset >= pos + data.size()) { // The requested offset is at or beyond the end of the file. This can // happen if `offset` is not block-aligned, and the read returns the last // block in the file, which does not extend all the way out to `offset`. std::stringstream os;
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Thu Jul 16 01:39:09 GMT 2020 - 11.1K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
CancellationManager* cancellation_manager_ TF_GUARDED_BY(execution_mutex_); // Outputs: std::vector<TensorHandlePtr> op_outputs_ TF_GUARDED_BY(execution_mutex_); // TF_Status is an incomplete type and so can't be stack allocated. To avoid // unnecessary allocations each Execute call, we keep one heap-allocated // version for the thread. StatusPtr status_ TF_GUARDED_BY(execution_mutex_); const std::string device_;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/eager/c_api_distributed_test.cc
} return tensorflow::errors::Internal("Graph pass runs for more than once!"); } private: bool first_call_ = true; }; // After the graph pass is registered, it takes effect globally and can affect // other test cases. Define a static variable to switch it on and off. bool GraphErrorInjectionPass::enabled_ = false; // Test to ensure that a registered graph optimization pass is only executed
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 23.5K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/modular_filesystem_registration.cc
// TensorFlow will know to not call these on behalf of users; // * increased security as plugins will not be able to alter function table // after loading up. Thus, malicious plugins can't alter functionality to // probe for gadgets inside core TensorFlow. We can even protect the area // of memory where the copies reside to not allow any more writes to it // after all copies are created. template <typename T>
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Mar 07 22:08:43 GMT 2023 - 12.8K bytes - Viewed (0) -
tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache_test.cc
ReadCache(&cache, "a", i * block_size, block_size, &out)); std::vector<char> x(block_size, 'x'); EXPECT_EQ(out, x); })); } // The `threads` destructor blocks until the threads can be joined, once their // respective reads finish (which happens once they are all concurrently being // executed, or 10 seconds have passed). } TEST(RamFileBlockCacheTest, CoalesceConcurrentReads) {
C++ - Registered: Tue Apr 23 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:57 GMT 2021 - 23.2K bytes - Viewed (0)