Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 14 for Cuda (0.02 sec)

  1. .bazelrc

    #     release_cpu_linux:               Toolchain and CUDA options for Linux CPU builds.
    #     release_gpu_linux:               Toolchain and CUDA options for Linux GPU builds.
    #     release_cpu_macos:               Toolchain and CUDA options for MacOS CPU builds.
    #     release_cpu_windows:             Toolchain and CUDA options for Windows CPU builds.
    
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Mon Oct 28 22:02:31 UTC 2024
    - 51.3K bytes
    - Viewed (0)
  2. ci/official/utilities/rename_and_verify_wheels.sh

      fi
    fi
    # VERY basic check to ensure the [and-cuda] package variant is installable.
    # Checks TFCI_BAZEL_COMMON_ARGS for "gpu" or "cuda", implying that the test is
    # relevant. All of the GPU test machines have CUDA installed via other means,
    # so I am not sure how to verify that the dependencies themselves are valid for
    # the moment.
    if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 02 21:18:17 UTC 2024
    - 4.3K bytes
    - Viewed (0)
  3. configure.py

        write_repo_env_to_bazelrc('cuda', env_var, local_path)
    
    
    def set_other_cuda_vars(environ_cp):
      """Set other CUDA related variables."""
      # If CUDA is enabled, always use GPU during build and test.
      if environ_cp.get('TF_CUDA_CLANG') == '1':
        write_to_bazelrc('build --config=cuda_clang')
      else:
        write_to_bazelrc('build --config=cuda')
    
    
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 02 22:16:02 UTC 2024
    - 48.2K bytes
    - Viewed (0)
  4. ci/official/containers/linux_arm64/Dockerfile

    # Install devtoolset devel dependencies
    COPY setup.sources.sh /setup.sources.sh
    COPY setup.packages.sh /setup.packages.sh
    COPY devel.packages.txt /devel.packages.txt
    COPY cuda.packages.txt /cuda.packages.txt
    RUN /setup.sources.sh && /setup.packages.sh /devel.packages.txt
    
    # Install various tools.
    # - bats: bash unit testing framework
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Fri Nov 01 08:30:33 UTC 2024
    - 4.4K bytes
    - Viewed (0)
  5. WORKSPACE

    tf_workspace0()
    
    load(
        "@local_tsl//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl",
        "cuda_json_init_repository",
    )
    
    cuda_json_init_repository()
    
    load(
        "@cuda_redist_json//:distributions.bzl",
        "CUDA_REDISTRIBUTIONS",
        "CUDNN_REDISTRIBUTIONS",
    )
    load(
        "@local_tsl//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",
        "cuda_redist_init_repositories",
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Fri Oct 11 16:49:28 UTC 2024
    - 3K bytes
    - Viewed (0)
  6. ci/official/utilities/code_check_full.bats

        done < $BATS_TEST_TMPDIR/missing_deps
        exit 1
      fi
    }
    
    # The Python package is not allowed to depend on any CUDA packages.
    @test "Pip package doesn't depend on CUDA" {
      bazel cquery \
        --experimental_cc_shared_library \
        --@local_config_cuda//:enable_cuda \
        --@local_config_cuda//cuda:include_cuda_libs=false \
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 23 18:48:35 UTC 2024
    - 13.6K bytes
    - Viewed (0)
  7. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 16 05:28:35 UTC 2024
    - 53.5K bytes
    - Viewed (0)
  8. CONTRIBUTING.md

            flag.
    
            ```bash
            export flags="--config=opt --config=cuda -k"
            ```
    
        *   For TensorFlow versions prior v.2.18.0: Add CUDA paths to
            LD_LIBRARY_PATH and add the `cuda` option flag.
    
            ```bash
            export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Wed Oct 23 06:20:12 UTC 2024
    - 15.9K bytes
    - Viewed (0)
  9. ci/official/containers/ml_build/README.md

    WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA).
    
    This container branches off from
    /tensorflow/tools/tf_sig_build_dockerfiles/. However, since
    hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the
    requirements installed on the original container can be removed to reduce the
    footprint of the container and make it more reusable across different ML
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Tue Sep 24 20:45:58 UTC 2024
    - 416 bytes
    - Viewed (0)
  10. ci/official/wheel.sh

    # limitations under the License.
    # ==============================================================================
    source "${BASH_SOURCE%/*}/utilities/setup.sh"
    
    # Record GPU count and CUDA version status
    if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then
      tfrun nvidia-smi
    fi
    
    # Update the version numbers for Nightly only
    if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
    Registered: Tue Nov 05 12:39:12 UTC 2024
    - Last Modified: Mon Oct 14 23:45:36 UTC 2024
    - 2.2K bytes
    - Viewed (0)
Back to top