- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 19 for CUDA (0.08 sec)
-
ci/official/containers/linux_arm64/cuda.packages.txt
# CuDNN: https://docs.nvidia.com/deeplearning/sdk/cudnn-install/index.html#ubuntu-network-installation libcudnn9-dev-cuda-12=9.1.1.17-1 libcudnn9-cuda-12=9.1.1.17-1 # This can be removed once NVIDIA publishes a cuda-12.3.2 Docker image. # For now it ensures that we install at least version 12.3.107 of PTXAS, # since 12.3.103 has a bug.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri May 24 17:22:37 UTC 2024 - 366 bytes - Viewed (0) -
.bazelrc
# release_cpu_linux: Toolchain and CUDA options for Linux CPU builds. # release_gpu_linux: Toolchain and CUDA options for Linux GPU builds. # release_cpu_macos: Toolchain and CUDA options for MacOS CPU builds. # release_cpu_windows: Toolchain and CUDA options for Windows CPU builds.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 28 22:02:31 UTC 2024 - 51.3K bytes - Viewed (0) -
.github/bot_config.yml
**1. Installing **TensorFlow-GPU** (TF) prebuilt binaries** Make sure you are using compatible TF and CUDA versions. Please refer following TF version and CUDA version compatibility table. | TF | CUDA | | :-------------: | :-------------: | | 2.5.0 | 11.2 | | 2.4.0 | 11.0 | | 2.1.0 - 2.3.0 | 10.1 |
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Jul 15 05:00:54 UTC 2024 - 4K bytes - Viewed (0) -
ci/official/utilities/rename_and_verify_wheels.sh
fi fi # VERY basic check to ensure the [and-cuda] package variant is installable. # Checks TFCI_BAZEL_COMMON_ARGS for "gpu" or "cuda", implying that the test is # relevant. All of the GPU test machines have CUDA installed via other means, # so I am not sure how to verify that the dependencies themselves are valid for # the moment. if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 02 21:18:17 UTC 2024 - 4.3K bytes - Viewed (0) -
configure.py
write_repo_env_to_bazelrc('cuda', env_var, local_path) def set_other_cuda_vars(environ_cp): """Set other CUDA related variables.""" # If CUDA is enabled, always use GPU during build and test. if environ_cp.get('TF_CUDA_CLANG') == '1': write_to_bazelrc('build --config=cuda_clang') else: write_to_bazelrc('build --config=cuda')
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 02 22:16:02 UTC 2024 - 48.2K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/code_check_full.bats
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Aug 14 18:47:44 UTC 2024 - 12.9K bytes - Viewed (0) -
ci/official/containers/linux_arm64/Dockerfile
# Install devtoolset devel dependencies COPY setup.sources.sh /setup.sources.sh COPY setup.packages.sh /setup.packages.sh COPY devel.packages.txt /devel.packages.txt COPY cuda.packages.txt /cuda.packages.txt RUN /setup.sources.sh && /setup.packages.sh /devel.packages.txt # Install various tools. # - bats: bash unit testing framework
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Nov 01 08:30:33 UTC 2024 - 4.4K bytes - Viewed (0) -
WORKSPACE
tf_workspace0() load( "@local_tsl//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl", "cuda_json_init_repository", ) cuda_json_init_repository() load( "@cuda_redist_json//:distributions.bzl", "CUDA_REDISTRIBUTIONS", "CUDNN_REDISTRIBUTIONS", ) load( "@local_tsl//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl", "cuda_redist_init_repositories",
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Oct 11 16:49:28 UTC 2024 - 3K bytes - Viewed (0) -
ci/official/utilities/code_check_full.bats
done < $BATS_TEST_TMPDIR/missing_deps exit 1 fi } # The Python package is not allowed to depend on any CUDA packages. @test "Pip package doesn't depend on CUDA" { bazel cquery \ --experimental_cc_shared_library \ --@local_config_cuda//:enable_cuda \ --@local_config_cuda//cuda:include_cuda_libs=false \
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 23 18:48:35 UTC 2024 - 13.6K bytes - Viewed (0) -
tensorflow/BUILD
# Config setting that is satisfied when TensorFlow is being built with CUDA # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS). alias( name = "is_cuda_enabled", actual = if_oss( "@local_config_cuda//:is_cuda_enabled", "@local_config_cuda//cuda:using_config_cuda", ), ) # Config setting that is satisfied when CUDA device code should be compiled
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 16 05:28:35 UTC 2024 - 53.5K bytes - Viewed (0)