Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 37 for CUDA (0.01 sec)

  1. .bazelrc

    #     release_cpu_linux:               Toolchain and CUDA options for Linux CPU builds.
    #     release_gpu_linux:               Toolchain and CUDA options for Linux GPU builds.
    #     release_cpu_macos:               Toolchain and CUDA options for MacOS CPU builds.
    #     release_cpu_windows:             Toolchain and CUDA options for Windows CPU builds.
    # LINT.IfChange
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Fri Dec 26 23:20:26 UTC 2025
    - 56.8K bytes
    - Viewed (0)
  2. configure.py

        write_repo_env_to_bazelrc('cuda', env_var, local_path)
    
    
    def set_other_cuda_vars(environ_cp):
      """Set other CUDA related variables."""
      # If CUDA is enabled, always use GPU during build and test.
      if environ_cp.get('TF_CUDA_CLANG') == '1':
        write_to_bazelrc('build --config=cuda_clang')
      else:
        write_to_bazelrc('build --config=cuda')
    
    
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Wed Apr 30 15:18:54 UTC 2025
    - 48.3K bytes
    - Viewed (0)
  3. ci/official/utilities/code_check_full.bats

        done < $BATS_TEST_TMPDIR/missing_deps
        exit 1
      fi
    }
    
    # The Python package is not allowed to depend on any CUDA packages.
    @test "Pip package doesn't depend on CUDA" {
      bazel cquery \
        --experimental_cc_shared_library \
        --@local_config_cuda//:enable_cuda \
        --@local_config_cuda//cuda:include_cuda_libs=false \
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Fri Dec 19 18:47:57 UTC 2025
    - 13.5K bytes
    - Viewed (0)
  4. ci/official/utilities/rename_and_verify_wheels.sh

        "$python" -m pip install numpy==1.26.4
      else
        "$python" -m pip install numpy==1.26.0
      fi
    fi
    if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then
      echo "Checking to make sure tensorflow[and-cuda] is installable..."
      "$python" -m pip install "$(echo *.whl)[and-cuda]" $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS
    else
      "$python" -m pip install *.whl $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS
    fi
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Mon Sep 22 21:39:32 UTC 2025
    - 4.4K bytes
    - Viewed (0)
  5. ci/official/containers/ml_build/setup.sources.cudnn.sh

    export DEBIAN_FRONTEND=noninteractive
    
    # Fetch the NVIDIA key.
    apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub;
    
    # Set up sources for NVIDIA CUDNN.
    cat >/etc/apt/sources.list.d/nvidia.list <<SOURCES
    # NVIDIA
    deb https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /
    
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Tue Feb 18 20:42:21 UTC 2025
    - 1.2K bytes
    - Viewed (0)
  6. CONTRIBUTING.md

            flag.
    
            ```bash
            export flags="--config=linux --config=cuda -k"
            ```
    
        *   For TensorFlow versions prior v.2.18.0: Add CUDA paths to
            LD_LIBRARY_PATH and add the `cuda` option flag.
    
            ```bash
            export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Sat Jan 11 04:47:59 UTC 2025
    - 15.9K bytes
    - Viewed (0)
  7. ci/official/containers/ml_build/Dockerfile

    RUN ln -sf /usr/bin/python3.12 /usr/bin/python
    RUN ln -sf /usr/lib/python3.12 /usr/lib/tf_python
    
    # Link the compat driver to the location if available.
    RUN if [ -e "/usr/local/cuda/compat/libcuda.so.1" ]; then ln -s /usr/local/cuda/compat/libcuda.so.1 /usr/lib/x86_64-linux-gnu/libcuda.so.1; fi
    
    # Install various tools.
    # - bats: bash unit testing framework
    # - bazelisk: always use the correct bazel version
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Thu Dec 18 21:55:23 UTC 2025
    - 4.5K bytes
    - Viewed (0)
  8. ci/official/containers/ml_build/README.md

    WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA).
    
    This container branches off from
    /tensorflow/tools/tf_sig_build_dockerfiles/. However, since
    hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the
    requirements installed on the original container can be removed to reduce the
    footprint of the container and make it more reusable across different ML
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Tue Sep 24 20:45:58 UTC 2024
    - 416 bytes
    - Viewed (0)
  9. .github/ISSUE_TEMPLATE/tflite-other.md

    -   type: input id: Compiler attributes: label: GCC/Compiler version
        description: if compiling from source placeholder: validations: required:
        false
    
    -   type: input id: Cuda attributes: label: CUDA/cuDNN version description:
        placeholder: validations: required: false
    
    -   type: input id: Gpu attributes: label: GPU model and memory description: if
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Thu Dec 29 22:28:29 UTC 2022
    - 3.4K bytes
    - Viewed (0)
  10. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Registered: Tue Dec 30 12:39:10 UTC 2025
    - Last Modified: Wed Nov 12 19:21:56 UTC 2025
    - 53.1K bytes
    - Viewed (0)
Back to top