- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 13 for GPUs (0.02 seconds)
-
WORKSPACE
load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl", "cuda_json_init_repository", ) cuda_json_init_repository() load( "@cuda_redist_json//:distributions.bzl", "CUDA_REDISTRIBUTIONS", "CUDNN_REDISTRIBUTIONS", ) load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 5.1K bytes - Click Count (0) -
ci/official/envs/linux_x86_cuda
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ARGS="--gpus all" TFCI_LIB_SUFFIX="-gpu-linux-x86_64" # TODO: Set back to 610M once the wheel size is fixed.
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Tue Feb 18 22:52:46 GMT 2025 - 1.1K bytes - Click Count (0) -
SECURITY.md
### Hardware attacks Physical GPUs or TPUs can also be the target of attacks. [Published research](https://scholar.google.com/scholar?q=gpu+side+channel) shows that it might be possible to use side channel attacks on the GPU to leak data from other running models or processes in the same system. GPUs can also have implementation bugs that might allow attackers to leave malicious code running
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Wed Oct 16 16:10:43 GMT 2024 - 9.6K bytes - Click Count (0) -
CITATION.cff
represent computation, shared state, and the operations that mutate that state. It maps the nodes of a dataflow graph across many machines in a cluster, and within a machine across multiple computational devices, including multicore CPUs, general purpose GPUs, and custom-designed ASICs known as Tensor Processing Units (TPUs). This architecture gives flexibility to the application developer, whereas in previous “parameter server” designs the management of shared state is built into the system, TensorFlow...
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Sep 06 15:26:23 GMT 2021 - 3.5K bytes - Click Count (0) -
configure.py
Args: environ_cp: copy of the os.environ. var_name: string for name of environment variable, e.g. "TF_NEED_CUDA". query_item: string for feature related to the variable, e.g. "CUDA for Nvidia GPUs". enabled_by_default: boolean for default behavior. question: optional string for how to ask for user input. yes_reply: optional string for reply when feature is enabled.Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Wed Apr 30 15:18:54 GMT 2025 - 48.3K bytes - Click Count (0) -
.bazelrc
# See https://developer.nvidia.com/cuda-gpus#compute # `compute_XY` enables PTX embedding in addition to SASS. PTX # is forward compatible beyond the current compute capability major # release while SASS is only forward compatible inside the current # major release. Example: sm_80 kernels can run on sm_89 GPUs but # not on sm_90 GPUs. compute_80 kernels though can also run on sm_90 GPUs.
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 56.8K bytes - Click Count (0) -
tensorflow/c/c_test_util.cc
TF_SetAttrType(desc, "T", TF_INT32); // Set device to CPU since there is no version of split for int32 on GPU // TODO(iga): Convert all these helpers and tests to use floats because // they are usually available on GPUs. After doing this, remove TF_SetDevice // call in c_api_function_test.cc TF_SetDevice(desc, "/cpu:0"); *op = TF_FinishOperation(desc, s); ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s);
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Oct 04 05:55:32 GMT 2025 - 17.8K bytes - Click Count (1) -
RELEASE.md
on Ampere based GPUs.TensorFloat-32, or TF32 for short, is a math mode for NVIDIA Ampere based GPUs which causes certain float32 ops, such as matrix multiplications and convolutions, to run much faster on Ampere GPUs but with reduced precision. This reduced precision has not beenCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Tue Oct 28 22:27:41 GMT 2025 - 740.4K bytes - Click Count (3) -
CHANGELOG/CHANGELOG-1.3.md
* Use local disk for ConfigMap volume instead of tmpfs ([#25306](https://github.com/kubernetes/kubernetes/pull/25306), [@pmorie](https://github.com/pmorie)) * Alpha support for scheduling pods on machines with NVIDIA GPUs whose kubelets use the `--experimental-nvidia-gpus` flag, using the alpha.kubernetes.io/nvidia-gpu resource ([#24836](https://github.com/kubernetes/kubernetes/pull/24836), [@therc](https://github.com/therc))
Created: Fri Dec 26 09:05:12 GMT 2025 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 84K bytes - Click Count (0) -
CHANGELOG/CHANGELOG-1.7.md
* Create clusters with GPUs in GCE by specifying `type=<gpu-type>,count=<gpu-count>` to NODE_ACCELERATORS environment variable. ([#45130](https://github.com/kubernetes/kubernetes/pull/45130), [@vishh](https://github.com/vishh)) * List of available GPUs - [https://cloud.google.com/compute/docs/gpus/#introduction](https://cloud.google.com/compute/docs/gpus/#introduction)Created: Fri Dec 26 09:05:12 GMT 2025 - Last Modified: Thu May 05 13:44:43 GMT 2022 - 308.7K bytes - Click Count (1)