- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 18 for NVIDIA (0.09 sec)
-
ci/official/containers/linux_arm64/cuda.packages.txt
# CuDNN: https://docs.nvidia.com/deeplearning/sdk/cudnn-install/index.html#ubuntu-network-installation libcudnn9-dev-cuda-12=9.1.1.17-1 libcudnn9-cuda-12=9.1.1.17-1 # This can be removed once NVIDIA publishes a cuda-12.3.2 Docker image. # For now it ensures that we install at least version 12.3.107 of PTXAS, # since 12.3.103 has a bug.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 24 17:22:37 UTC 2024 - 366 bytes - Viewed (0) -
tensorflow/compiler/jit/tests/BUILD
srcs = [ "device_compiler_serialize_test.cc", ], tags = [ "config-cuda-only", "no_oss", # This test only runs with GPU. "requires-gpu-nvidia", "xla", ], deps = [ ":device_compiler_test_helper", "//tensorflow/compiler/jit:compilation_passes", "//tensorflow/compiler/jit:flags", "//tensorflow/core:test",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 27 18:00:18 UTC 2024 - 3.7K bytes - Viewed (0) -
ci/official/README.md
# container and start fresh, run "docker rm -f tf". Removing the container # destroys some temporary bazel data and causes longer builds. # # You will need the NVIDIA Container Toolkit for GPU testing: # https://github.com/NVIDIA/nvidia-container-toolkit # # Note: if you interrupt a bazel command on docker (ctrl-c), you # will need to run `docker exec tf pkill bazel` to quit bazel. #
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 01 03:21:19 UTC 2024 - 8K bytes - Viewed (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/build-arm64:tf-2-16-multi-python TFCI_DOCKER_PULL_ENABLE=1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 03 16:53:33 UTC 2024 - 1.5K bytes - Viewed (0) -
ci/official/libtensorflow.sh
# ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then tfrun python3 tensorflow/tools/ci_build/update_version.py --nightly fi
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jan 19 19:07:48 UTC 2024 - 1.5K bytes - Viewed (0) -
ci/official/wheel.sh
# ============================================================================== source "${BASH_SOURCE%/*}/utilities/setup.sh" # Record GPU count and CUDA version status if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then tfrun nvidia-smi fi # Update the version numbers for Nightly only if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then tfrun python3 tensorflow/tools/ci_build/update_version.py --nightly
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 06 21:54:13 UTC 2024 - 1.8K bytes - Viewed (0) -
ci/official/containers/linux_arm64/Dockerfile
COPY builder.patchelf/build_patchelf.sh /build_patchelf.sh COPY apt.conf /etc/apt/ RUN /build_patchelf.sh ################################################################################ FROM nvidia/cuda:12.3.1-devel-ubuntu20.04 as devel ################################################################################ COPY --from=builder /dt10 /dt10
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jan 08 09:32:19 UTC 2024 - 4.1K bytes - Viewed (0) -
configure.py
Args: environ_cp: copy of the os.environ. var_name: string for name of environment variable, e.g. "TF_NEED_CUDA". query_item: string for feature related to the variable, e.g. "CUDA for Nvidia GPUs". enabled_by_default: boolean for default behavior. question: optional string for how to ask for user input. yes_reply: optional string for reply when feature is enabled.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 04:32:44 UTC 2024 - 53.8K bytes - Viewed (0) -
docs/en/data/external_links.yml
Neon link: https://neon.tech/blog/deploy-a-serverless-fastapi-app-with-neon-postgres-and-aws-app-runner-at-any-scale title: Deploy a Serverless FastAPI App with Neon Postgres and AWS App Runner at any scale - author: Kurtis Pykes - NVIDIA link: https://developer.nvidia.com/blog/building-a-machine-learning-microservice-with-fastapi/ title: Building a Machine Learning Microservice with FastAPI - author: Ravgeet Dhillon - Twilio link: https://www.twilio.com/en-us/blog/booking-appointments-twilio-notion-fastapi...
Registered: Mon Jun 17 08:32:26 UTC 2024 - Last Modified: Wed Jun 12 00:47:57 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/compiler/jit/BUILD
tf_cuda_cc_test( name = "pjrt_compile_util_test", srcs = ["pjrt_compile_util_test.cc"], tags = [ "config-cuda-only", "no_oss", # This test only runs with GPU. "requires-gpu-nvidia", "xla", ], deps = [ ":pjrt_compile_util", ":test_util", ":xla_gpu_jit", "//tensorflow/cc:function_ops", "//tensorflow/cc:math_ops",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 61.5K bytes - Viewed (0)