Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for GPU (0.06 sec)

  1. .bazelrc

    # CUDA WHEEL
    test:linux_cuda_wheel_test_filters --test_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310,-no_oss_py313
    test:linux_cuda_wheel_test_filters --build_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310,-no_oss_py313
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Fri Aug 22 21:03:34 UTC 2025
    - 56K bytes
    - Viewed (0)
  2. tensorflow/c/c_api_test.cc

    TEST(CAPI, Session_Min_GPU) {
      const string gpu_device = GPUDeviceName();
      // Skip this test if no GPU is available.
      if (gpu_device.empty()) return;
    
      RunMinTest(gpu_device, /*use_XLA=*/false);
    }
    
    TEST(CAPI, Session_Min_XLA_GPU) {
      const string gpu_device = GPUDeviceName();
      // Skip this test if no GPU is available.
      if (gpu_device.empty()) return;
    
      RunMinTest(gpu_device, /*use_XLA=*/true);
    }
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Fri Dec 27 12:18:10 UTC 2024
    - 97K bytes
    - Viewed (0)
  3. CHANGELOG/CHANGELOG-1.3.md

    * Do not query the metadata server to find out if running on GCE.  Retry metadata server query for gcr if running on gce. ([#28871](https://github.com/kubernetes/kubernetes/pull/28871), [@vishh](https://github.com/vishh))
    * Fix GPU resource validation ([#28743](https://github.com/kubernetes/kubernetes/pull/28743), [@therc](https://github.com/therc))
    Registered: Fri Sep 05 09:05:11 UTC 2025
    - Last Modified: Thu Dec 24 02:28:26 UTC 2020
    - 84K bytes
    - Viewed (0)
  4. tensorflow/BUILD

    )
    
    config_setting(
        name = "with_xla_support",
        define_values = {"with_xla_support": "true"},
        visibility = ["//visibility:public"],
    )
    
    # By default, XLA GPU is compiled into tensorflow when building with
    # --config=cuda even when `with_xla_support` is false. The config setting
    # here allows us to override the behavior if needed.
    config_setting(
        name = "no_xla_deps_in_cuda",
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Thu Aug 28 19:11:51 UTC 2025
    - 53.4K bytes
    - Viewed (0)
  5. tensorflow/c/c_api_function_test.cc

        for (auto input : inputs) {
          TF_AddInput(desc, input);
        }
        // Set device to CPU because some ops inside the function might not be
        // available on GPU.
        TF_SetDevice(desc, "/cpu:0");
        *op = TF_FinishOperation(desc, s_);
        ASSERT_EQ(TF_OK, TF_GetCode(s_)) << TF_Message(s_);
        ASSERT_NE(*op, nullptr);
      }
    
      FunctionDef fdef() {
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Thu Jul 20 22:08:54 UTC 2023
    - 63.6K bytes
    - Viewed (1)
Back to top