- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 29 for Cuda (0.06 sec)
-
tensorflow/BUILD
# Config setting that is satisfied when TensorFlow is being built with CUDA # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS). alias( name = "is_cuda_enabled", actual = if_oss( "@local_config_cuda//:is_cuda_enabled", "@local_config_cuda//cuda:using_config_cuda", ), ) # Config setting that is satisfied when CUDA device code should be compiled
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 16 05:28:35 UTC 2024 - 53.5K bytes - Viewed (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
description: If compiling from source - type: input id: Compiler attributes: label: GCC/compiler version description: If compiling from source - type: input id: Cuda attributes: label: CUDA/cuDNN version - type: input id: Gpu attributes: label: GPU model and memory description: If compiling from source - type: textarea id: what-happened
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Jun 28 18:25:42 UTC 2023 - 3.7K bytes - Viewed (0) -
CONTRIBUTING.md
flag. ```bash export flags="--config=opt --config=cuda -k" ``` * For TensorFlow versions prior v.2.18.0: Add CUDA paths to LD_LIBRARY_PATH and add the `cuda` option flag. ```bash export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 23 06:20:12 UTC 2024 - 15.9K bytes - Viewed (0) -
ci/official/README.md
These "env" files match up with an environment matrix that roughly covers: - Different Python versions - Linux, MacOS, and Windows machines (these pool definitions are internal) - x86 and arm64 - CPU-only, or with NVIDIA CUDA support (Linux only), or with TPUs ## How to Test Your Changes to TensorFlow You may check how your changes will affect TensorFlow by: 1. Creating a PR and observing the presubmit test results
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Thu Feb 01 03:21:19 UTC 2024 - 8K bytes - Viewed (0) -
ci/official/libtensorflow.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Jan 19 19:07:48 UTC 2024 - 1.5K bytes - Viewed (0) -
ci/official/containers/ml_build/README.md
WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA). This container branches off from /tensorflow/tools/tf_sig_build_dockerfiles/. However, since hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the requirements installed on the original container can be removed to reduce the footprint of the container and make it more reusable across different ML
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Sep 24 20:45:58 UTC 2024 - 416 bytes - Viewed (0) -
ci/official/wheel.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 14 23:45:36 UTC 2024 - 2.2K bytes - Viewed (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/build-arm64:tf-2-18-multi-python TFCI_DOCKER_PULL_ENABLE=1
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 14 23:45:36 UTC 2024 - 1.5K bytes - Viewed (0) -
RELEASE.md
* Hermetic CUDA support is added. Hermetic CUDA uses a specific downloadable version of CUDA instead of the user’s locally installed CUDA. Bazel will download CUDA, CUDNN and NCCL distributions, and then use CUDA libraries and tools as dependencies in various Bazel targets. This enables more reproducible builds for Google ML projects and supported CUDA versions.
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Oct 22 14:33:53 UTC 2024 - 735.3K bytes - Viewed (0) -
ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc
# Change the value of CACHEBUSTER when upgrading the toolchain, or when testing # different compilation methods. E.g. for a PR to test a new CUDA version, set # the CACHEBUSTER to the PR number. build --action_env=CACHEBUSTER=20220325 # Use Python 3.X as installed in container image build --action_env PYTHON_BIN_PATH="/usr/local/bin/python3"
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Fri Jul 12 20:16:57 UTC 2024 - 5.7K bytes - Viewed (0)