- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 35 for CUDA (0.04 sec)
-
configure.py
write_repo_env_to_bazelrc('cuda', env_var, local_path) def set_other_cuda_vars(environ_cp): """Set other CUDA related variables.""" # If CUDA is enabled, always use GPU during build and test. if environ_cp.get('TF_CUDA_CLANG') == '1': write_to_bazelrc('build --config=cuda_clang') else: write_to_bazelrc('build --config=cuda')
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Wed Apr 30 15:18:54 UTC 2025 - 48.3K bytes - Viewed (0) -
CONTRIBUTING.md
flag. ```bash export flags="--config=linux --config=cuda -k" ``` * For TensorFlow versions prior v.2.18.0: Add CUDA paths to LD_LIBRARY_PATH and add the `cuda` option flag. ```bash export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Sat Jan 11 04:47:59 UTC 2025 - 15.9K bytes - Viewed (0) -
ci/official/utilities/code_check_full.bats
done < $BATS_TEST_TMPDIR/missing_deps exit 1 fi } # The Python package is not allowed to depend on any CUDA packages. @test "Pip package doesn't depend on CUDA" { bazel cquery \ --experimental_cc_shared_library \ --@local_config_cuda//:enable_cuda \ --@local_config_cuda//cuda:include_cuda_libs=false \
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Wed Aug 06 20:43:08 UTC 2025 - 13.4K bytes - Viewed (0) -
ci/official/containers/ml_build/setup.sources.cudnn.sh
export DEBIAN_FRONTEND=noninteractive # Fetch the NVIDIA key. apt-key adv --fetch-keys https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/3bf863cc.pub; # Set up sources for NVIDIA CUDNN. cat >/etc/apt/sources.list.d/nvidia.list <<SOURCES # NVIDIA deb https://developer.download.nvidia.com/compute/cuda/repos/ubuntu2204/x86_64/ /
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Tue Feb 18 20:42:21 UTC 2025 - 1.2K bytes - Viewed (0) -
ci/official/containers/ml_build/Dockerfile
# Make sure clang is on the path RUN ln -s /usr/lib/llvm-18/bin/clang /usr/bin/clang # Link the compat driver to the location if available. RUN if [ -e "/usr/local/cuda/compat/libcuda.so.1" ]; then ln -s /usr/local/cuda/compat/libcuda.so.1 /usr/lib/x86_64-linux-gnu/libcuda.so.1; fi # Install various tools. # - bats: bash unit testing framework # - bazelisk: always use the correct bazel version
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Fri Aug 01 02:44:57 UTC 2025 - 4.9K bytes - Viewed (0) -
ci/official/containers/ml_build/README.md
WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA). This container branches off from /tensorflow/tools/tf_sig_build_dockerfiles/. However, since hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the requirements installed on the original container can be removed to reduce the footprint of the container and make it more reusable across different ML
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Tue Sep 24 20:45:58 UTC 2024 - 416 bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tflite-other.md
- type: input id: Compiler attributes: label: GCC/Compiler version description: if compiling from source placeholder: validations: required: false - type: input id: Cuda attributes: label: CUDA/cuDNN version description: placeholder: validations: required: false - type: input id: Gpu attributes: label: GPU model and memory description: if
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Thu Dec 29 22:28:29 UTC 2022 - 3.4K bytes - Viewed (0) -
tensorflow/BUILD
# Config setting that is satisfied when TensorFlow is being built with CUDA # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS). alias( name = "is_cuda_enabled", actual = if_oss( "@local_config_cuda//:is_cuda_enabled", "@local_config_cuda//cuda:using_config_cuda", ), ) # Config setting that is satisfied when CUDA device code should be compiled
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Thu Aug 28 19:11:51 UTC 2025 - 53.4K bytes - Viewed (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-docker.pkg.dev/ml-oss-artifacts-published/ml-public-container/ml-build-arm64:latest
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Wed Jun 04 01:09:09 UTC 2025 - 1.6K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
description: If compiling from source - type: input id: Compiler attributes: label: GCC/compiler version description: If compiling from source - type: input id: Cuda attributes: label: CUDA/cuDNN version - type: input id: Gpu attributes: label: GPU model and memory description: If compiling from source - type: textarea id: what-happened
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Wed Jun 28 18:25:42 UTC 2023 - 3.7K bytes - Viewed (0)