- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 13 for GPUs (0.07 seconds)
-
ci/official/envs/linux_x86_cuda
TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow" TFCI_DOCKER_ARGS="--gpus all" TFCI_LIB_SUFFIX="-gpu-linux-x86_64" # TODO: Set back to 610M once the wheel size is fixed.
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Tue Feb 18 22:52:46 GMT 2025 - 1.1K bytes - Click Count (0) -
SECURITY.md
### Hardware attacks Physical GPUs or TPUs can also be the target of attacks. [Published research](https://scholar.google.com/scholar?q=gpu+side+channel) shows that it might be possible to use side channel attacks on the GPU to leak data from other running models or processes in the same system. GPUs can also have implementation bugs that might allow attackers to leave malicious code running
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Wed Oct 16 16:10:43 GMT 2024 - 9.6K bytes - Click Count (0) -
CITATION.cff
represent computation, shared state, and the operations that mutate that state. It maps the nodes of a dataflow graph across many machines in a cluster, and within a machine across multiple computational devices, including multicore CPUs, general purpose GPUs, and custom-designed ASICs known as Tensor Processing Units (TPUs). This architecture gives flexibility to the application developer, whereas in previous “parameter server” designs the management of shared state is built into the system, TensorFlow...
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Sep 06 15:26:23 GMT 2021 - 3.5K bytes - Click Count (0) -
configure.py
Args: environ_cp: copy of the os.environ. var_name: string for name of environment variable, e.g. "TF_NEED_CUDA". query_item: string for feature related to the variable, e.g. "CUDA for Nvidia GPUs". enabled_by_default: boolean for default behavior. question: optional string for how to ask for user input. yes_reply: optional string for reply when feature is enabled.Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Fri Dec 19 16:32:04 GMT 2025 - 48.3K bytes - Click Count (0) -
tensorflow/c/eager/c_api_test_util.h
tensorflow::ServerDef GetServerDef(int num_tasks); // Create a multi-client ServerDef with the given `job_name`, add `num_tasks` // tasks and `num_virtual_gpus` virtual GPUs in it. tensorflow::ServerDef GetMultiClientServerDef(const std::string& job_name, int num_tasks, int num_virtual_gpus = 0);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Jul 17 23:43:59 GMT 2023 - 7.7K bytes - Click Count (0) -
.bazelrc
# See https://developer.nvidia.com/cuda-gpus#compute # `compute_XY` enables PTX embedding in addition to SASS. PTX # is forward compatible beyond the current compute capability major # release while SASS is only forward compatible inside the current # major release. Example: sm_80 kernels can run on sm_89 GPUs but # not on sm_90 GPUs. compute_80 kernels though can also run on sm_90 GPUs.
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Mar 28 04:33:01 GMT 2026 - 58.9K bytes - Click Count (0) -
tensorflow/c/c_test_util.cc
TF_SetAttrType(desc, "T", TF_INT32); // Set device to CPU since there is no version of split for int32 on GPU // TODO(iga): Convert all these helpers and tests to use floats because // they are usually available on GPUs. After doing this, remove TF_SetDevice // call in c_api_function_test.cc TF_SetDevice(desc, "/cpu:0"); *op = TF_FinishOperation(desc, s); ASSERT_EQ(TF_OK, TF_GetCode(s)) << TF_Message(s);
Created: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Sat Oct 04 05:55:32 GMT 2025 - 17.8K bytes - Click Count (1) -
RELEASE.md
on Ampere based GPUs.TensorFloat-32, or TF32 for short, is a math mode for NVIDIA Ampere based GPUs which causes certain float32 ops, such as matrix multiplications and convolutions, to run much faster on Ampere GPUs but with reduced precision. This reduced precision has not beenCreated: Tue Apr 07 12:39:13 GMT 2026 - Last Modified: Mon Mar 30 18:31:38 GMT 2026 - 746.5K bytes - Click Count (3) -
CHANGELOG/CHANGELOG-1.6.md
resource availability for critical system pods. * **[alpha]** `--experimental-nvidia-gpus` flag is replaced by `Accelerators` alpha feature gate along with addition of support for multiple Nvidia GPUs. - To use GPUs, pass `Accelerators=true` as part of `--feature-gates` flag. - More information [here](https://vishh.github.io/docs/user-guide/gpus/). * A pod’s Quality of Service Class is now available in its Status.
Created: Fri Apr 03 09:05:14 GMT 2026 - Last Modified: Thu Dec 24 02:28:26 GMT 2020 - 304K bytes - Click Count (0) -
CHANGELOG/CHANGELOG-1.7.md
* Create clusters with GPUs in GCE by specifying `type=<gpu-type>,count=<gpu-count>` to NODE_ACCELERATORS environment variable. ([#45130](https://github.com/kubernetes/kubernetes/pull/45130), [@vishh](https://github.com/vishh)) * List of available GPUs - [https://cloud.google.com/compute/docs/gpus/#introduction](https://cloud.google.com/compute/docs/gpus/#introduction)Created: Fri Apr 03 09:05:14 GMT 2026 - Last Modified: Thu May 05 13:44:43 GMT 2022 - 308.7K bytes - Click Count (1)