Search Options

Display Count
Sort
Preferred Language
Advanced Search

Results 1 - 10 of 13 for GPUs (0.03 seconds)

  1. WORKSPACE

    load(
        "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl",
        "cuda_json_init_repository",
    )
    
    cuda_json_init_repository()
    
    load(
        "@cuda_redist_json//:distributions.bzl",
        "CUDA_REDISTRIBUTIONS",
        "CUDNN_REDISTRIBUTIONS",
    )
    load(
        "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Fri Dec 26 23:20:26 GMT 2025
    - 5.1K bytes
    - Click Count (0)
  2. ci/official/envs/linux_x86_cuda

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda
    TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ARGS="--gpus all"
    TFCI_LIB_SUFFIX="-gpu-linux-x86_64"
    # TODO: Set back to 610M once the wheel size is fixed.
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Tue Feb 18 22:52:46 GMT 2025
    - 1.1K bytes
    - Click Count (0)
  3. SECURITY.md

    ### Hardware attacks
    
    Physical GPUs or TPUs can also be the target of attacks. [Published
    research](https://scholar.google.com/scholar?q=gpu+side+channel) shows that it
    might be possible to use side channel attacks on the GPU to leak data from other
    running models or processes in the same system. GPUs can also have
    implementation bugs that might allow attackers to leave malicious code running
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Wed Oct 16 16:10:43 GMT 2024
    - 9.6K bytes
    - Click Count (0)
  4. CITATION.cff

    represent computation, shared state, and the operations that mutate that state. It maps the nodes of a dataflow graph across many machines in a cluster, and within a machine across multiple computational devices, including multicore CPUs, general purpose GPUs, and custom-designed ASICs known as Tensor Processing Units (TPUs). This architecture gives flexibility to the application developer, whereas in previous “parameter server” designs the management of shared state is built into the system, TensorFlow...
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Mon Sep 06 15:26:23 GMT 2021
    - 3.5K bytes
    - Click Count (0)
  5. configure.py

      Args:
        environ_cp: copy of the os.environ.
        var_name: string for name of environment variable, e.g. "TF_NEED_CUDA".
        query_item: string for feature related to the variable, e.g. "CUDA for
          Nvidia GPUs".
        enabled_by_default: boolean for default behavior.
        question: optional string for how to ask for user input.
        yes_reply: optional string for reply when feature is enabled.
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Wed Apr 30 15:18:54 GMT 2025
    - 48.3K bytes
    - Click Count (0)
  6. .bazelrc

    # See https://developer.nvidia.com/cuda-gpus#compute
    # `compute_XY` enables PTX embedding in addition to SASS. PTX
    # is forward compatible beyond the current compute capability major
    # release while SASS is only forward compatible inside the current
    # major release. Example: sm_80 kernels can run on sm_89 GPUs but
    # not on sm_90 GPUs. compute_80 kernels though can also run on sm_90 GPUs.
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Fri Dec 26 23:20:26 GMT 2025
    - 56.8K bytes
    - Click Count (0)
  7. tensorflow/c/c_test_util.cc

      TF_SetAttrType(desc, "T", TF_INT32);
      // Set device to CPU since there is no version of split for int32 on GPU
      // TODO(iga): Convert all these helpers and tests to use floats because
      // they are usually available on GPUs. After doing this, remove TF_SetDevice
      // call in c_api_function_test.cc
      TF_SetDevice(desc, "/cpu:0");
      *op = TF_FinishOperation(desc, s);
      ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s);
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Sat Oct 04 05:55:32 GMT 2025
    - 17.8K bytes
    - Click Count (1)
  8. RELEASE.md

            on Ampere based GPUs.TensorFloat-32, or TF32 for short, is a math mode
            for NVIDIA Ampere based GPUs which causes certain float32 ops, such as
            matrix multiplications and convolutions, to run much faster on Ampere
            GPUs but with reduced precision. This reduced precision has not been
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Tue Oct 28 22:27:41 GMT 2025
    - 740.4K bytes
    - Click Count (3)
  9. CHANGELOG/CHANGELOG-1.3.md

    * Use local disk for ConfigMap volume instead of tmpfs ([#25306](https://github.com/kubernetes/kubernetes/pull/25306), [@pmorie](https://github.com/pmorie))
    * Alpha support for scheduling pods on machines with NVIDIA GPUs whose kubelets use the `--experimental-nvidia-gpus` flag, using the alpha.kubernetes.io/nvidia-gpu resource  ([#24836](https://github.com/kubernetes/kubernetes/pull/24836), [@therc](https://github.com/therc))
    Created: Fri Dec 26 09:05:12 GMT 2025
    - Last Modified: Thu Dec 24 02:28:26 GMT 2020
    - 84K bytes
    - Click Count (0)
  10. CHANGELOG/CHANGELOG-1.7.md

      * Create clusters with GPUs in GCE by specifying `type=<gpu-type>,count=<gpu-count>` to NODE_ACCELERATORS environment variable. ([#45130](https://github.com/kubernetes/kubernetes/pull/45130), [@vishh](https://github.com/vishh))
    
        * List of available GPUs - [https://cloud.google.com/compute/docs/gpus/#introduction](https://cloud.google.com/compute/docs/gpus/#introduction)
    
    Created: Fri Dec 26 09:05:12 GMT 2025
    - Last Modified: Thu May 05 13:44:43 GMT 2022
    - 308.7K bytes
    - Click Count (1)
Back to Top