- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 210 for cpui (0.02 sec)
-
docs/de/docs/deployment/concepts.md
Wie viele Systemressourcen möchten Sie verbrauchen/nutzen? Sie mögen „nicht viel“ denken, aber in Wirklichkeit möchten Sie tatsächlich **so viel wie möglich ohne Absturz** verwenden.
Registered: Sun Dec 28 07:19:09 UTC 2025 - Last Modified: Tue Dec 02 17:32:56 UTC 2025 - 21.5K bytes - Viewed (0) -
docs/en/docs/deployment/concepts.md
On the other hand, if you have 2 servers and you are using **100% of their CPU and RAM**, at some point one process will ask for more memory, and the server will have to use the disk as "memory" (which can be thousands of times slower), or even **crash**. Or one process might need to do some computation and would have to wait until the CPU is free again.
Registered: Sun Dec 28 07:19:09 UTC 2025 - Last Modified: Sun Aug 31 09:15:41 UTC 2025 - 18.6K bytes - Viewed (1) -
docs/en/docs/deployment/server-workers.md
## Recap { #recap } You can use multiple worker processes with the `--workers` CLI option with the `fastapi` or `uvicorn` commands to take advantage of **multi-core CPUs**, to run **multiple processes in parallel**. You could use these tools and ideas if you are setting up **your own deployment system** while taking care of the other deployment concepts yourself.Registered: Sun Dec 28 07:19:09 UTC 2025 - Last Modified: Sun Aug 31 09:15:41 UTC 2025 - 8.3K bytes - Viewed (0) -
docs/pt/docs/deployment/server-workers.md
## Recapitular { #recap } Você pode usar vários processos de trabalho com a opção CLI `--workers` com os comandos `fastapi` ou `uvicorn` para aproveitar as vantagens de **CPUs multi-core** e executar **vários processos em paralelo**. Você pode usar essas ferramentas e ideias se estiver configurando **seu próprio sistema de implantação** enquanto cuida dos outros conceitos de implantação.Registered: Sun Dec 28 07:19:09 UTC 2025 - Last Modified: Wed Nov 12 16:23:57 UTC 2025 - 8.6K bytes - Viewed (0) -
docs/metrics/prometheus/list.md
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Tue Aug 12 18:20:36 UTC 2025 - 43.4K bytes - Viewed (0) -
cmd/metrics-resource.go
Registered: Sun Dec 28 19:28:13 UTC 2025 - Last Modified: Fri Oct 10 18:57:03 UTC 2025 - 17.2K bytes - Viewed (0) -
docs/de/docs/deployment/server-workers.md
## Zusammenfassung { #recap } Sie können mehrere Workerprozesse mit der `--workers`-CLI-Option über die `fastapi`- oder `uvicorn`-Befehle nutzen, um **Multikern-CPUs** auszunutzen und **mehrere Prozesse parallel** auszuführen. Sie könnten diese Tools und Ideen nutzen, wenn Sie **Ihr eigenes Deployment-System** einrichten und sich dabei selbst um die anderen Deployment-Konzepte kümmern.Registered: Sun Dec 28 07:19:09 UTC 2025 - Last Modified: Tue Dec 02 17:32:56 UTC 2025 - 8.8K bytes - Viewed (0) -
.github/bot_config.yml
Therefore on any CPU that does not have these instruction sets, either CPU or GPU version of TF will fail to load. Apparently, your CPU model does not support AVX instruction sets. You can still use TensorFlow with the alternatives given below: * Try Google Colab to use TensorFlow.Registered: Tue Dec 30 12:39:10 UTC 2025 - Last Modified: Mon Jun 30 16:38:59 UTC 2025 - 4K bytes - Viewed (1) -
RELEASE.md
* `tf.raw_ops.Bucketize` op on CPU. * `tf.where` op for data types `tf.int32`/`tf.uint32`/`tf.int8`/`tf.uint8`/`tf.int64`. * `tf.random.normal` op for output data type `tf.float32` on CPU. * `tf.random.uniform` op for output data type `tf.float32` on CPU. * `tf.random.categorical` op for output data type `tf.int64` on CPU. * `tensorflow.experimental.tensorrt`:Registered: Tue Dec 30 12:39:10 UTC 2025 - Last Modified: Tue Oct 28 22:27:41 UTC 2025 - 740.4K bytes - Viewed (3) -
tensorflow/c/README.md
- Nightly builds: - [Linux CPU-only](https://storage.googleapis.com/tensorflow-nightly/github/tensorflow/lib_package/libtensorflow-cpu-linux-x86_64.tar.gz) - [Linux GPU](https://storage.googleapis.com/tensorflow-nightly/github/tensorflow/lib_package/libtensorflow-gpu-linux-x86_64.tar.gz)
Registered: Tue Dec 30 12:39:10 UTC 2025 - Last Modified: Tue Oct 23 01:38:30 UTC 2018 - 539 bytes - Viewed (0)