- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 12 for NVIDIA (0.11 sec)
-
ci/official/containers/linux_arm64/cuda.packages.txt
# CuDNN: https://docs.nvidia.com/deeplearning/sdk/cudnn-install/index.html#ubuntu-network-installation libcudnn9-dev-cuda-12=9.1.1.17-1 libcudnn9-cuda-12=9.1.1.17-1 # This can be removed once NVIDIA publishes a cuda-12.3.2 Docker image. # For now it ensures that we install at least version 12.3.107 of PTXAS, # since 12.3.103 has a bug.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 24 17:22:37 UTC 2024 - 366 bytes - Viewed (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ENABLE=1 TFCI_DOCKER_IMAGE=us-central1-docker.pkg.dev/tensorflow-sigs/tensorflow/build-arm64:tf-2-16-multi-python TFCI_DOCKER_PULL_ENABLE=1
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 03 16:53:33 UTC 2024 - 1.5K bytes - Viewed (0) -
configure.py
Args: environ_cp: copy of the os.environ. var_name: string for name of environment variable, e.g. "TF_NEED_CUDA". query_item: string for feature related to the variable, e.g. "CUDA for Nvidia GPUs". enabled_by_default: boolean for default behavior. question: optional string for how to ask for user input. yes_reply: optional string for reply when feature is enabled.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Jun 10 04:32:44 UTC 2024 - 53.8K bytes - Viewed (0) -
docs/en/data/external_links.yml
Neon link: https://neon.tech/blog/deploy-a-serverless-fastapi-app-with-neon-postgres-and-aws-app-runner-at-any-scale title: Deploy a Serverless FastAPI App with Neon Postgres and AWS App Runner at any scale - author: Kurtis Pykes - NVIDIA link: https://developer.nvidia.com/blog/building-a-machine-learning-microservice-with-fastapi/ title: Building a Machine Learning Microservice with FastAPI - author: Ravgeet Dhillon - Twilio link: https://www.twilio.com/en-us/blog/booking-appointments-twilio-notion-fastapi...
Registered: Mon Jun 17 08:32:26 UTC 2024 - Last Modified: Wed Jun 12 00:47:57 UTC 2024 - 22K bytes - Viewed (0) -
tensorflow/compiler/jit/BUILD
tf_cuda_cc_test( name = "pjrt_compile_util_test", srcs = ["pjrt_compile_util_test.cc"], tags = [ "config-cuda-only", "no_oss", # This test only runs with GPU. "requires-gpu-nvidia", "xla", ], deps = [ ":pjrt_compile_util", ":test_util", ":xla_gpu_jit", "//tensorflow/cc:function_ops", "//tensorflow/cc:math_ops",
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 31 00:41:19 UTC 2024 - 61.5K bytes - Viewed (0) -
RELEASE.md
The `tensorflow` pip package has a new, optional installation method for Linux that installs necessary Nvidia CUDA libraries through pip. As long as the Nvidia driver is already installed on the system, you may now run `pip install tensorflow[and-cuda]` to install TensorFlow's Nvidia CUDA library dependencies in the Python environment. Aside from the Nvidia driver, no other pre-existing Nvidia CUDA packages are necessary. * Enable JIT-compiled i64-indexed kernels on GPU for large tensors...
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 730.3K bytes - Viewed (0) -
.bazelrc
build:cuda_clang --config=cuda # Enable TensorRT optimizations https://developer.nvidia.com/tensorrt build:cuda_clang --config=tensorrt build:cuda_clang --action_env=TF_CUDA_CLANG="1" build:cuda_clang --@local_config_cuda//:cuda_compiler=clang # Select supported compute capabilities (supported graphics cards). # This is the same as the official TensorFlow builds. # See https://developer.nvidia.com/cuda-gpus#compute
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 17:12:54 UTC 2024 - 52.9K bytes - Viewed (0) -
src/debug/elf/elf.go
EM_TILEPRO Machine = 188 /* Tilera TILEPro multicore architecture family */ EM_MICROBLAZE Machine = 189 /* Xilinx MicroBlaze 32-bit RISC soft processor core */ EM_CUDA Machine = 190 /* NVIDIA CUDA architecture */ EM_TILEGX Machine = 191 /* Tilera TILE-Gx multicore architecture family */ EM_CLOUDSHIELD Machine = 192 /* CloudShield architecture family */
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 16 00:01:16 UTC 2024 - 134.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops_a_m.cc
if (one_by_one && trivial_strides && trivial_dilations) { return "NHWC"; } // If filter spatial dimensions are unknown or not 1x1 we prefer NCHW, because // it's the fastest option on NVIDIA GPUs with cuDNN library support. return "NCHW"; } //===----------------------------------------------------------------------===// // Conv2dBackpropFilterOp
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 146.7K bytes - Viewed (0) -
cluster/gce/gci/configure-helper.sh
fi if [[ "${ENABLE_NVIDIA_GPU_DEVICE_PLUGIN:-}" == "true" ]]; then setup-addon-manifests "addons" "device-plugins/nvidia-gpu" fi # Setting up the konnectivity-agent daemonset if [[ "${RUN_KONNECTIVITY_PODS:-false}" == "true" ]]; then setup-addon-manifests "addons" "konnectivity-agent" setup-konnectivity-agent-manifest
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon Jun 10 22:07:47 UTC 2024 - 141.1K bytes - Viewed (0)