Search Options

Display Count
Sort
Preferred Language
Advanced Search

Results 1 - 10 of 24 for CUDA (0.02 seconds)

  1. ci/official/containers/ml_build/rbe_nvidia.packages.txt

    # The RBE machine itself has older kernel mode driver, and it requires
    # nvidia driver to be installed.
    nvidia-driver-580-open
    # TODO(b/445248346): The Docker image shouldn't have cuda-compat installed.
    # However, hermetic CUDA forward-compatibility mode is still missing some
    # libraries.
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Thu Sep 18 00:19:40 GMT 2025
    - 307 bytes
    - Click Count (0)
  2. ci/official/requirements_updater/nvidia-requirements.txt

    nvidia-cublas-cu12>=12.5.3.2,<13.0
    nvidia-cuda-cupti-cu12>=12.5.82,<13.0
    nvidia-cuda-nvcc-cu12>=12.5.82,<13.0
    nvidia-cuda-nvrtc-cu12>=12.5.82,<13.0
    nvidia-cuda-runtime-cu12>=12.5.82,<13.0
    # The upper bound is set for the CUDNN API compatibility.
    # See
    # https://docs.nvidia.com/deeplearning/cudnn/backend/latest/developer/forward-compatibility.html#cudnn-api-compatibility
    nvidia-cudnn-cu12>=9.3.0.75,<10.0
    nvidia-cufft-cu12>=11.2.3.61,<12.0
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Wed Sep 03 23:57:17 GMT 2025
    - 646 bytes
    - Click Count (0)
  3. .bazelrc

    #     release_cpu_linux:               Toolchain and CUDA options for Linux CPU builds.
    #     release_gpu_linux:               Toolchain and CUDA options for Linux GPU builds.
    #     release_cpu_macos:               Toolchain and CUDA options for MacOS CPU builds.
    #     release_cpu_windows:             Toolchain and CUDA options for Windows CPU builds.
    # LINT.IfChange
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Sat Mar 28 04:33:01 GMT 2026
    - 58.9K bytes
    - Click Count (0)
  4. .github/bot_config.yml

       
       **1. Installing **TensorFlow-GPU** (TF) prebuilt binaries**
       
       
       Make sure you are using compatible TF and CUDA versions.
       Please refer following TF version and CUDA version compatibility table.
       
       | TF  | CUDA |
       
       | :-------------: | :-------------: |
       
       | 2.5.0  | 11.2 |
       
       | 2.4.0  | 11.0 |
       
       | 2.1.0 - 2.3.0  | 10.1 |
       
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Mon Jun 30 16:38:59 GMT 2025
    - 4K bytes
    - Click Count (1)
  5. configure.py

        write_repo_env_to_bazelrc('cuda', env_var, local_path)
    
    
    def set_other_cuda_vars(environ_cp):
      """Set other CUDA related variables."""
      # If CUDA is enabled, always use GPU during build and test.
      if environ_cp.get('TF_CUDA_CLANG') == '1':
        write_to_bazelrc('build --config=cuda_clang')
      else:
        write_to_bazelrc('build --config=cuda')
    
    
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Fri Dec 19 16:32:04 GMT 2025
    - 48.3K bytes
    - Click Count (0)
  6. ci/official/utilities/code_check_full.bats

        done < $BATS_TEST_TMPDIR/missing_deps
        exit 1
      fi
    }
    
    # The Python package is not allowed to depend on any CUDA packages.
    @test "Pip package doesn't depend on CUDA" {
      bazel cquery \
        --experimental_cc_shared_library \
        --@local_config_cuda//:enable_cuda \
        --@local_config_cuda//cuda:include_cuda_libs=false \
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Wed Jan 28 22:41:17 GMT 2026
    - 13.6K bytes
    - Click Count (0)
  7. ci/official/containers/ml_build/Dockerfile

    RUN ln -sf /usr/bin/python3.12 /usr/bin/python
    RUN ln -sf /usr/lib/python3.12 /usr/lib/tf_python
    
    # Link the compat driver to the location if available.
    RUN if [ -e "/usr/local/cuda/compat/libcuda.so.1" ]; then ln -s /usr/local/cuda/compat/libcuda.so.1 /usr/lib/x86_64-linux-gnu/libcuda.so.1; fi
    
    # Install various tools.
    # - bats: bash unit testing framework
    # - bazelisk: always use the correct bazel version
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Fri Mar 20 15:35:12 GMT 2026
    - 4.5K bytes
    - Click Count (0)
  8. ci/official/utilities/rename_and_verify_wheels.sh

        "$python" -m pip install numpy==1.26.4
      else
        "$python" -m pip install numpy==1.26.0
      fi
    fi
    if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then
      echo "Checking to make sure tensorflow[and-cuda] is installable..."
      "$python" -m pip install "$(echo *.whl)[and-cuda]" $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS
    else
      "$python" -m pip install *.whl $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS
    fi
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Mon Sep 22 21:39:32 GMT 2025
    - 4.4K bytes
    - Click Count (0)
  9. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Tue Mar 24 21:00:18 GMT 2026
    - 53.1K bytes
    - Click Count (0)
  10. ci/official/envs/linux_arm64

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64
    # Note: this is not set to "--cpu", because that changes the package name
    # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow"
    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=us-docker.pkg.dev/ml-oss-artifacts-published/ml-public-container/ml-build-arm64:latest
    Created: Tue Apr 07 12:39:13 GMT 2026
    - Last Modified: Thu Mar 12 16:45:37 GMT 2026
    - 1.6K bytes
    - Click Count (0)
Back to Top