- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 19 for gpu (0.04 sec)
-
.bazelrc
# CUDA WHEEL test:linux_cuda_wheel_test_filters --test_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310 test:linux_cuda_wheel_test_filters --build_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 28 22:02:31 UTC 2024 - 51.3K bytes - Viewed (0) -
ci/official/utilities/rename_and_verify_wheels.sh
# VERY basic check to ensure the [and-cuda] package variant is installable. # Checks TFCI_BAZEL_COMMON_ARGS for "gpu" or "cuda", implying that the test is # relevant. All of the GPU test machines have CUDA installed via other means, # so I am not sure how to verify that the dependencies themselves are valid for # the moment. if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then echo "Checking to make sure tensorflow[and-cuda] is installable..."
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 02 21:18:17 UTC 2024 - 4.3K bytes - Viewed (0) -
CONTRIBUTING.md
and [GPU developer Dockerfile](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/tools/dockerfiles/dockerfiles/devel-gpu.Dockerfile) for the required packages. Alternatively, use the said [tensorflow/build Docker images](https://hub.docker.com/r/tensorflow/build) (`tensorflow/tensorflow:devel` and `tensorflow/tensorflow:devel-gpu` are no
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 23 06:20:12 UTC 2024 - 15.9K bytes - Viewed (0) -
ci/official/envs/linux_x86_cuda_build
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ARGS="--gpus all" TFCI_LIB_SUFFIX="-gpu-linux-x86_64"
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 28 17:57:41 UTC 2024 - 1.1K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 05:11:17 UTC 2024 - 12.9K bytes - Viewed (0) -
RELEASE.md
`XNNPACK` delegate automatically when the model has a `fp32` operation. * GPU * Allow GPU acceleration starting with internal graph nodes * Experimental support for quantized models with the Android GPU delegate * Add GPU delegate whitelist. * Rename GPU whitelist -> compatibility (list). * Improve GPU compatibility list entries from crash reports. * NNAPI * Set default value for
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Tue Oct 22 14:33:53 UTC 2024 - 735.3K bytes - Viewed (0) -
ci/official/envs/linux_x86_cuda
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ARGS="--gpus all" TFCI_LIB_SUFFIX="-gpu-linux-x86_64"
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 14 23:45:36 UTC 2024 - 1K bytes - Viewed (0) -
SECURITY.md
### Hardware attacks Physical GPUs or TPUs can also be the target of attacks. [Published research](https://scholar.google.com/scholar?q=gpu+side+channel) shows that it might be possible to use side channel attacks on the GPU to leak data from other running models or processes in the same system. GPUs can also have implementation bugs that might allow attackers to leave malicious code running
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Wed Oct 16 16:10:43 UTC 2024 - 9.6K bytes - Viewed (0) -
ci/official/wheel.sh
# limitations under the License. # ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Mon Oct 14 23:45:36 UTC 2024 - 2.2K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.cc
// threadpool of GPU event mgr, as that can trigger more callbacks to be // scheduled on that same threadpool, causing a deadlock in cases where the // caller of event_mgr->ThenExecute() blocks on the completion of the callback // (as in the case of ConstOp kernel creation on GPU, which involves copying a // CPU tensor to GPU).
Registered: Tue Nov 05 12:39:12 UTC 2024 - Last Modified: Sat Oct 12 16:27:48 UTC 2024 - 29.5K bytes - Viewed (0)