Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 35 for CUDA (0.02 sec)

  1. configure.py

        write_repo_env_to_bazelrc('cuda', env_var, local_path)
    
    
    def set_other_cuda_vars(environ_cp):
      """Set other CUDA related variables."""
      # If CUDA is enabled, always use GPU during build and test.
      if environ_cp.get('TF_CUDA_CLANG') == '1':
        write_to_bazelrc('build --config=cuda_clang')
      else:
        write_to_bazelrc('build --config=cuda')
    
    
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Wed Apr 30 15:18:54 UTC 2025
    - 48.3K bytes
    - Viewed (0)
  2. CONTRIBUTING.md

            flag.
    
            ```bash
            export flags="--config=linux --config=cuda -k"
            ```
    
        *   For TensorFlow versions prior v.2.18.0: Add CUDA paths to
            LD_LIBRARY_PATH and add the `cuda` option flag.
    
            ```bash
            export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Sat Jan 11 04:47:59 UTC 2025
    - 15.9K bytes
    - Viewed (0)
  3. ci/official/utilities/code_check_full.bats

        done < $BATS_TEST_TMPDIR/missing_deps
        exit 1
      fi
    }
    
    # The Python package is not allowed to depend on any CUDA packages.
    @test "Pip package doesn't depend on CUDA" {
      bazel cquery \
        --experimental_cc_shared_library \
        --@local_config_cuda//:enable_cuda \
        --@local_config_cuda//cuda:include_cuda_libs=false \
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Wed Aug 06 20:43:08 UTC 2025
    - 13.4K bytes
    - Viewed (0)
  4. ci/official/containers/ml_build/setup.sources.cudnn.sh

    export DEBIAN_FRONTEND=noninteractive
    
    # Fetch the NVIDIA key.
    apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub;
    
    # Set up sources for NVIDIA CUDNN.
    cat >/etc/apt/sources.list.d/nvidia.list <<SOURCES
    # NVIDIA
    deb https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /
    
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Tue Feb 18 20:42:21 UTC 2025
    - 1.2K bytes
    - Viewed (0)
  5. ci/official/containers/ml_build/Dockerfile

    # Make sure clang is on the path
    RUN ln -s /usr/lib/llvm-18/bin/clang /usr/bin/clang
    
    # Link the compat driver to the location if available.
    RUN if [ -e "/usr/local/cuda/compat/libcuda.so.1" ]; then ln -s /usr/local/cuda/compat/libcuda.so.1 /usr/lib/x86_64-linux-gnu/libcuda.so.1; fi
    
    # Install various tools.
    # - bats: bash unit testing framework
    # - bazelisk: always use the correct bazel version
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Fri Aug 01 02:44:57 UTC 2025
    - 4.9K bytes
    - Viewed (0)
  6. ci/official/containers/ml_build/README.md

    WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA).
    
    This container branches off from
    /tensorflow/tools/tf_sig_build_dockerfiles/. However, since
    hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the
    requirements installed on the original container can be removed to reduce the
    footprint of the container and make it more reusable across different ML
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Tue Sep 24 20:45:58 UTC 2024
    - 416 bytes
    - Viewed (0)
  7. .github/ISSUE_TEMPLATE/tflite-other.md

    -   type: input id: Compiler attributes: label: GCC/Compiler version
        description: if compiling from source placeholder: validations: required:
        false
    
    -   type: input id: Cuda attributes: label: CUDA/cuDNN version description:
        placeholder: validations: required: false
    
    -   type: input id: Gpu attributes: label: GPU model and memory description: if
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Thu Dec 29 22:28:29 UTC 2022
    - 3.4K bytes
    - Viewed (0)
  8. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Thu Aug 28 19:11:51 UTC 2025
    - 53.4K bytes
    - Viewed (0)
  9. ci/official/envs/linux_arm64

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64
    # Note: this is not set to "--cpu", because that changes the package name
    # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow"
    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=us-docker.pkg.dev/ml-oss-artifacts-published/ml-public-container/ml-build-arm64:latest
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Wed Jun 04 01:09:09 UTC 2025
    - 1.6K bytes
    - Viewed (0)
  10. .github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml

          description: If compiling from source
      - type: input
        id: Compiler
        attributes:
          label: GCC/compiler version
          description: If compiling from source
      - type: input
        id: Cuda
        attributes:
          label: CUDA/cuDNN version
      - type: input
        id: Gpu
        attributes:
          label: GPU model and memory
          description: If compiling from source
      - type: textarea
        id: what-happened
    Registered: Tue Sep 09 12:39:10 UTC 2025
    - Last Modified: Wed Jun 28 18:25:42 UTC 2023
    - 3.7K bytes
    - Viewed (0)
Back to top