- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 20 for GPU (0.02 seconds)
-
.github/bot_config.yml
cuda_comment: > From the template it looks like you are installing **TensorFlow** (TF) prebuilt binaries: * For TF-GPU - See point 1 * For TF-CPU - See point 2 ----------------------------------------------------------------------------------------------- **1. Installing **TensorFlow-GPU** (TF) prebuilt binaries** Make sure you are using compatible TF and CUDA versions.Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Jun 30 16:38:59 GMT 2025 - 4K bytes - Click Count (1) -
.bazelrc
# CUDA WHEEL test:linux_cuda_wheel_test_filters --test_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310,-no_oss_py313 test:linux_cuda_wheel_test_filters --build_tag_filters=gpu,requires-gpu,-no_gpu,-no_oss,-tf_tosa,-oss_excluded,-oss_serial,-benchmark-test,-no_cuda11,-no_oss_py38,-no_oss_py39,-no_oss_py310,-no_oss_py313
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 56.8K bytes - Click Count (0) -
README.md
[pip package](https://www.tensorflow.org/install/pip), to [enable GPU support](https://www.tensorflow.org/install/gpu), use a [Docker container](https://www.tensorflow.org/install/docker), and [build from source](https://www.tensorflow.org/install/source). To install the current release, which includes support for [CUDA-enabled GPU cards](https://www.tensorflow.org/install/gpu) *(Ubuntu and Windows)*: ``` $ pip install tensorflow ```
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Jul 18 14:09:03 GMT 2025 - 11.6K bytes - Click Count (0) -
configure.py
(environ_cp.get('TF_NEED_ROCM', None) == '1')): test_and_build_filters += ['-no_windows_gpu', '-no_gpu'] else: test_and_build_filters.append('-gpu') elif is_macos(): test_and_build_filters += ['-gpu', '-nomac', '-no_mac', '-mac_excluded'] elif is_linux(): if ((environ_cp.get('TF_NEED_CUDA', None) == '1') or (environ_cp.get('TF_NEED_ROCM', None) == '1')):
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Wed Apr 30 15:18:54 GMT 2025 - 48.3K bytes - Click Count (0) -
RELEASE.md
`XNNPACK` delegate automatically when the model has a `fp32` operation. * GPU * Allow GPU acceleration starting with internal graph nodes * Experimental support for quantized models with the Android GPU delegate * Add GPU delegate whitelist. * Rename GPU whitelist -> compatibility (list). * Improve GPU compatibility list entries from crash reports. * NNAPI * Set default value forCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Tue Oct 28 22:27:41 GMT 2025 - 740.4K bytes - Click Count (3) -
tensorflow/c/c_api_experimental.cc
// threadpool of GPU event mgr, as that can trigger more callbacks to be // scheduled on that same threadpool, causing a deadlock in cases where the // caller of event_mgr->ThenExecute() blocks on the completion of the callback // (as in the case of ConstOp kernel creation on GPU, which involves copying a // CPU tensor to GPU).
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Oct 04 05:55:32 GMT 2025 - 29.4K bytes - Click Count (0) -
ci/official/utilities/rename_and_verify_wheels.sh
if [[ "$TFCI_PYTHON_VERSION" == "3.13" ]]; then "$python" -m pip install numpy==1.26.4 else "$python" -m pip install numpy==1.26.0 fi fi if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then echo "Checking to make sure tensorflow[and-cuda] is installable..." "$python" -m pip install "$(echo *.whl)[and-cuda]" $TFCI_PYTHON_VERIFY_PIP_INSTALL_ARGS else
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Sep 22 21:39:32 GMT 2025 - 4.4K bytes - Click Count (0) -
.github/workflows/build.yml
with: api-level: ${{ matrix.api-level }} arch: ${{ matrix.api-level == '34' && 'x86_64' || 'x86' }} force-avd-creation: false emulator-options: -no-window -gpu swiftshader_indirect -noaudio -no-boot-anim -camera-back none disable-animations: false script: echo "Generated AVD snapshot for caching." - name: Run TestsCreated: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Fri Dec 12 04:49:37 GMT 2025 - 18.6K bytes - Click Count (0) -
docs/ru/docs/advanced/events.md
А затем сразу после `yield` мы выгружаем модель. Этот код будет выполнен после того, как приложение закончит обрабатывать запросы, непосредственно перед shutdown. Это может, например, освободить ресурсы, такие как память или GPU. /// tip | Совет `shutdown` произойдёт, когда вы останавливаете приложение. Возможно, вам нужно запустить новую версию, или вы просто устали от него. 🤷 /// ### Функция lifespan { #lifespan-function }Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 20:41:43 GMT 2025 - 12.7K bytes - Click Count (0) -
docs/en/docs/advanced/events.md
And then, right after the `yield`, we unload the model. This code will be executed **after** the application **finishes handling requests**, right before the *shutdown*. This could, for example, release resources like memory or a GPU. /// tip The `shutdown` would happen when you are **stopping** the application. Maybe you need to start a new version, or you just got tired of running it. 🤷 ///
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 20:41:43 GMT 2025 - 7.9K bytes - Click Count (0)