- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 39 for GPU (0.06 seconds)
-
WORKSPACE
) load( "@rules_ml_toolchain//gpu/cuda:cuda_configure.bzl", "cuda_configure", ) cuda_configure(name = "local_config_cuda") load( "@rules_ml_toolchain//gpu/nccl:nccl_redist_init_repository.bzl", "nccl_redist_init_repository", ) nccl_redist_init_repository() load( "@rules_ml_toolchain//gpu/nccl:nccl_configure.bzl", "nccl_configure", )Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Apr 02 01:32:13 GMT 2026 - 5.3K bytes - Click Count (0) -
.bazelrc
# CUDA WHEEL test:linux_cuda_wheel_test_filters --test_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py310,-no_oss_py313 test:linux_cuda_wheel_test_filters --build_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py310,-no_oss_py313
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Mar 28 04:33:01 GMT 2026 - 58.9K bytes - Click Count (0) -
.github/bot_config.yml
cuda_comment: > From the template it looks like you are installing **TensorFlow** (TF) prebuilt binaries: * For TF-GPU - See point 1 * For TF-CPU - See point 2 ----------------------------------------------------------------------------------------------- **1. Installing **TensorFlow-GPU** (TF) prebuilt binaries** Make sure you are using compatible TF and CUDA versions.Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Jun 30 16:38:59 GMT 2025 - 4K bytes - Click Count (1) -
README.md
[pip package](https://www.tensorflow.org/install/pip), to [enable GPU support](https://www.tensorflow.org/install/gpu), use a [Docker container](https://www.tensorflow.org/install/docker), and [build from source](https://www.tensorflow.org/install/source). To install the current release, which includes support for [CUDA-enabled GPU cards](https://www.tensorflow.org/install/gpu) *(Ubuntu and Windows)*: ``` pip install tensorflow ```
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Apr 02 10:38:57 GMT 2026 - 11.6K bytes - Click Count (0) -
CONTRIBUTING.md
and [GPU developer Dockerfile](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/tools/dockerfiles/dockerfiles/devel-gpu.Dockerfile) for the required packages. Alternatively, use the said [tensorflow/build Docker images](https://hub.docker.com/r/tensorflow/build) (`tensorflow/tensorflow:devel` and `tensorflow/tensorflow:devel-gpu` are noCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Jan 11 04:47:59 GMT 2025 - 15.9K bytes - Click Count (0) -
ci/official/envs/linux_x86_cuda
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ARGS="--gpus all" TFCI_LIB_SUFFIX="-gpu-linux-x86_64" # TODO: Set back to 610M once the wheel size is fixed.
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Tue Feb 18 22:52:46 GMT 2025 - 1.1K bytes - Click Count (0) -
tensorflow/c/eager/c_api_test.cc
TFE_Op* matmul = MatMulOp(ctx, m, m); // Disable the test if no GPU is present. string gpu_device_name; if (GetDeviceName(ctx, &gpu_device_name, "GPU")) { TFE_OpSetDevice(matmul, "GPU:0", status); ASSERT_TRUE(TF_GetCode(status) == TF_OK) << TF_Message(status); const char* device_name = TFE_OpGetDevice(matmul, status); ASSERT_TRUE(strstr(device_name, "GPU:0") != nullptr); TFE_OpSetDevice(matmul, "CPU:0", status);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Oct 09 05:56:18 GMT 2025 - 94.6K bytes - Click Count (0) -
configure.py
(environ_cp.get('TF_NEED_ROCM', None) == '1')): test_and_build_filters += ['-no_windows_gpu', '-no_gpu'] else: test_and_build_filters.append('-gpu') elif is_macos(): test_and_build_filters += ['-gpu', '-nomac', '-no_mac', '-mac_excluded'] elif is_linux(): if ((environ_cp.get('TF_NEED_CUDA', None) == '1') or (environ_cp.get('TF_NEED_ROCM', None) == '1')):
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Fri Dec 19 16:32:04 GMT 2025 - 48.3K bytes - Click Count (0) -
tensorflow/c/eager/dlpack.cc
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Thu Mar 13 23:41:52 GMT 2025 - 13K bytes - Click Count (0) -
RELEASE.md
`XNNPACK` delegate automatically when the model has a `fp32` operation. * GPU * Allow GPU acceleration starting with internal graph nodes * Experimental support for quantized models with the Android GPU delegate * Add GPU delegate whitelist. * Rename GPU whitelist -> compatibility (list). * Improve GPU compatibility list entries from crash reports. * NNAPI * Set default value forCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Mar 30 18:31:38 GMT 2026 - 746.5K bytes - Click Count (3)