- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 21 for CUDA (0.02 seconds)
-
ci/official/containers/ml_build/rbe_nvidia.packages.txt
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Sep 18 00:19:40 GMT 2025 - 307 bytes - Click Count (0) -
ci/official/requirements_updater/nvidia-requirements.txt
nvidia-cublas-cu12>=12.5.3.2,<13.0 nvidia-cuda-cupti-cu12>=12.5.82,<13.0 nvidia-cuda-nvcc-cu12>=12.5.82,<13.0 nvidia-cuda-nvrtc-cu12>=12.5.82,<13.0 nvidia-cuda-runtime-cu12>=12.5.82,<13.0 # The upper bound is set for the CUDNN API compatibility. # See # https://docs.nvidia.com/deeplearning/cudnn/backend/latest/developer/forward-compatibility.html#cudnn-api-compatibility nvidia-cudnn-cu12>=9.3.0.75,<10.0 nvidia-cufft-cu12>=11.2.3.61,<12.0
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Sep 03 23:57:17 GMT 2025 - 646 bytes - Click Count (0) -
.github/bot_config.yml
**1. Installing **TensorFlow-GPU** (TF) prebuilt binaries** Make sure you are using compatible TF and CUDA versions. Please refer following TF version and CUDA version compatibility table. | TF | CUDA | | :-------------: | :-------------: | | 2.5.0 | 11.2 | | 2.4.0 | 11.0 | | 2.1.0 - 2.3.0 | 10.1 |
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Jun 30 16:38:59 GMT 2025 - 4K bytes - Click Count (1) -
ci/official/containers/ml_build/setup.sources.cudnn.sh
export DEBIAN_FRONTEND=noninteractive # Fetch the NVIDIA key. apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub; # Set up sources for NVIDIA CUDNN. cat >/etc/apt/sources.list.d/nvidia.list <<SOURCES # NVIDIA deb https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Tue Feb 18 20:42:21 GMT 2025 - 1.2K bytes - Click Count (0) -
ci/official/containers/ml_build/Dockerfile
RUN ln -sf /usr/bin/python3.12 /usr/bin/python RUN ln -sf /usr/lib/python3.12 /usr/lib/tf_python # Link the compat driver to the location if available. RUN if [ -e "/usr/local/cuda/compat/libcuda.so.1" ]; then ln -s /usr/local/cuda/compat/libcuda.so.1 /usr/lib/x86_64-linux-gnu/libcuda.so.1; fi # Install various tools. # - bats: bash unit testing framework # - bazelisk: always use the correct bazel version
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Fri Mar 20 15:35:12 GMT 2026 - 4.5K bytes - Click Count (0) -
ci/official/containers/ml_build/README.md
WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA). This container branches off from /tensorflow/tools/tf_sig_build_dockerfiles/. However, since hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the requirements installed on the original container can be removed to reduce the footprint of the container and make it more reusable across different ML
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Tue Sep 24 20:45:58 GMT 2024 - 416 bytes - Click Count (0) -
ci/official/utilities/rename_and_verify_wheels.sh
"$python" -m pip install numpy==1.26.4 else "$python" -m pip install numpy==1.26.0 fi fi if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then echo "Checking to make sure tensorflow[and-cuda] is installable..." "$python" -m pip install "$(echo *.whl)[and-cuda]" $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS else "$python" -m pip install *.whl $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS fiCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Sep 22 21:39:32 GMT 2025 - 4.4K bytes - Click Count (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
description: If compiling from source - type: input id: Compiler attributes: label: GCC/compiler version description: If compiling from source - type: input id: Cuda attributes: label: CUDA/cuDNN version - type: input id: Gpu attributes: label: GPU model and memory description: If compiling from source - type: textarea id: what-happenedCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Jun 28 18:25:42 GMT 2023 - 3.7K bytes - Click Count (0) -
.github/ISSUE_TEMPLATE/tflite-other.md
- type: input id: Compiler attributes: label: GCC/Compiler version description: if compiling from source placeholder: validations: required: false - type: input id: Cuda attributes: label: CUDA/cuDNN version description: placeholder: validations: required: false - type: input id: Gpu attributes: label: GPU model and memory description: ifCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Dec 29 22:28:29 GMT 2022 - 3.4K bytes - Click Count (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-docker.pkg.dev/ml-oss-artifacts-published/ml-public-container/ml-build-arm64:latest
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Mar 12 16:45:37 GMT 2026 - 1.6K bytes - Click Count (0)