- Sort Score
- Num 10 results
- Language All
Results 1 - 6 of 6 for GPU (0.01 seconds)
-
.bazelrc
# CUDA WHEEL test:linux_cuda_wheel_test_filters --test_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py310,-no_oss_py313 test:linux_cuda_wheel_test_filters --build_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py310,-no_oss_py313
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Mar 28 04:33:01 GMT 2026 - 58.9K bytes - Click Count (0) -
tensorflow/c/eager/c_api_test.cc
TFE_Op* matmul = MatMulOp(ctx, m, m); // Disable the test if no GPU is present. string gpu_device_name; if (GetDeviceName(ctx, &gpu_device_name, "GPU")) { TFE_OpSetDevice(matmul, "GPU:0", status); ASSERT_TRUE(TF_GetCode(status) == TF_OK) << TF_Message(status); const char* device_name = TFE_OpGetDevice(matmul, status); ASSERT_TRUE(strstr(device_name, "GPU:0") != nullptr); TFE_OpSetDevice(matmul, "CPU:0", status);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 94.6K bytes - Click Count (0) -
tensorflow/c/c_api_test.cc
TEST(CAPI, Session_Min_GPU) { const std::string gpu_device = GPUDeviceName(); // Skip this test if no GPU is available. if (gpu_device.empty()) return; RunMinTest(gpu_device, /*use_XLA=*/false); } TEST(CAPI, Session_Min_XLA_GPU) { const std::string gpu_device = GPUDeviceName(); // Skip this test if no GPU is available. if (gpu_device.empty()) return; RunMinTest(gpu_device, /*use_XLA=*/true);Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Jan 07 04:56:09 GMT 2026 - 97.3K bytes - Click Count (0) -
CHANGELOG/CHANGELOG-1.3.md
* Do not query the metadata server to find out if running on GCE. Retry metadata server query for gcr if running on gce. ([#28871](https://github.com/kubernetes/kubernetes/pull/28871), [@vishh](https://github.com/vishh)) * Fix GPU resource validation ([#28743](https://github.com/kubernetes/kubernetes/pull/28743), [@therc](https://github.com/therc))
Created: Fri Apr 03 09:05:14 GMT 2026 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 84K bytes - Click Count (0) -
tensorflow/BUILD
) config_setting( name = "with_xla_support", define_values = {"with_xla_support": "true"}, visibility = ["//visibility:public"], ) # By default, XLA GPU is compiled into tensorflow when building with # --config=cuda even when `with_xla_support` is false. The config setting # here allows us to override the behavior if needed. config_setting( name = "no_xla_deps_in_cuda",Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Tue Mar 24 21:00:18 GMT 2026 - 53.1K bytes - Click Count (0) -
tensorflow/c/c_api_function_test.cc
for (auto input : inputs) { TF_AddInput(desc, input); } // Set device to CPU because some ops inside the function might not be // available on GPU. TF_SetDevice(desc, "/cpu:0"); *op = TF_FinishOperation(desc, s_); ASSERT_EQ(TF_OK, TF_GetCode(s_)) << TF_Message(s_); ASSERT_NE(*op, nullptr); } FunctionDef fdef() {
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Jan 07 04:56:09 GMT 2026 - 63.9K bytes - Click Count (1)