Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 104 for cuda (0.03 sec)

  1. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 16 05:28:35 UTC 2024
    - 53.5K bytes
    - Viewed (0)
  2. .github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml

          description: If compiling from source
      - type: input
        id: Compiler
        attributes:
          label: GCC/compiler version
          description: If compiling from source
      - type: input
        id: Cuda
        attributes:
          label: CUDA/cuDNN version
      - type: input
        id: Gpu
        attributes:
          label: GPU model and memory
          description: If compiling from source
      - type: textarea
        id: what-happened
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Jun 28 18:25:42 UTC 2023
    - 3.7K bytes
    - Viewed (0)
  3. CONTRIBUTING.md

            flag.
    
            ```bash
            export flags="--config=opt --config=cuda -k"
            ```
    
        *   For TensorFlow versions prior v.2.18.0: Add CUDA paths to
            LD_LIBRARY_PATH and add the `cuda` option flag.
    
            ```bash
            export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 23 06:20:12 UTC 2024
    - 15.9K bytes
    - Viewed (0)
  4. ci/official/README.md

    These "env" files match up with an environment matrix that roughly covers:
    
    -   Different Python versions
    -   Linux, MacOS, and Windows machines (these pool definitions are internal)
    -   x86 and arm64
    -   CPU-only, or with NVIDIA CUDA support (Linux only), or with TPUs
    
    ## How to Test Your Changes to TensorFlow
    
    You may check how your changes will affect TensorFlow by:
    
    1. Creating a PR and observing the presubmit test results
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Thu Feb 01 03:21:19 UTC 2024
    - 8K bytes
    - Viewed (0)
  5. ci/official/libtensorflow.sh

    # limitations under the License.
    # ==============================================================================
    source "${BASH_SOURCE%/*}/utilities/setup.sh"
    
    # Record GPU count and CUDA version status
    if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then
      tfrun nvidia-smi
    fi
    
    # Update the version numbers for Nightly only
    if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Fri Jan 19 19:07:48 UTC 2024
    - 1.5K bytes
    - Viewed (0)
  6. ci/official/containers/ml_build/README.md

    WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA).
    
    This container branches off from
    /tensorflow/tools/tf_sig_build_dockerfiles/. However, since
    hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the
    requirements installed on the original container can be removed to reduce the
    footprint of the container and make it more reusable across different ML
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Tue Sep 24 20:45:58 UTC 2024
    - 416 bytes
    - Viewed (0)
  7. ci/official/wheel.sh

    # limitations under the License.
    # ==============================================================================
    source "${BASH_SOURCE%/*}/utilities/setup.sh"
    
    # Record GPU count and CUDA version status
    if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then
      tfrun nvidia-smi
    fi
    
    # Update the version numbers for Nightly only
    if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Mon Oct 14 23:45:36 UTC 2024
    - 2.2K bytes
    - Viewed (0)
  8. ci/official/envs/linux_arm64

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64
    # Note: this is not set to "--cpu", because that changes the package name
    # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow"
    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/build-arm64:tf-2-18-multi-python
    TFCI_DOCKER_PULL_ENABLE=1
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Mon Oct 14 23:45:36 UTC 2024
    - 1.5K bytes
    - Viewed (0)
  9. RELEASE.md

    * Hermetic CUDA support is added.
    
      Hermetic CUDA uses a specific downloadable version of CUDA instead of the user’s locally installed CUDA. Bazel will download CUDA, CUDNN and NCCL distributions, and then use CUDA libraries and tools as dependencies in various Bazel targets. This enables more reproducible builds for Google ML projects and supported CUDA versions.
    
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Tue Oct 22 14:33:53 UTC 2024
    - 735.3K bytes
    - Viewed (0)
  10. ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc

    # Change the value of CACHEBUSTER when upgrading the toolchain, or when testing
    # different compilation methods. E.g. for a PR to test a new CUDA version, set
    # the CACHEBUSTER to the PR number.
    build --action_env=CACHEBUSTER=20220325
    
    # Use Python 3.X as installed in container image
    build --action_env PYTHON_BIN_PATH="/usr/local/bin/python3"
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Fri Jul 12 20:16:57 UTC 2024
    - 5.7K bytes
    - Viewed (0)
Back to top