- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 194 for cpui (0.13 sec)
-
docs/de/docs/deployment/concepts.md
Wie viele Systemressourcen möchten Sie verbrauchen/nutzen? Sie mögen „nicht viel“ denken, aber in Wirklichkeit möchten Sie tatsächlich **so viel wie möglich ohne Absturz** verwenden.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sun May 11 13:37:26 UTC 2025 - 20.6K bytes - Viewed (0) -
docs/metrics/v3.md
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Aug 12 18:20:36 UTC 2025 - 45.2K bytes - Viewed (0) -
docs/en/docs/deployment/concepts.md
On the other hand, if you have 2 servers and you are using **100% of their CPU and RAM**, at some point one process will ask for more memory, and the server will have to use the disk as "memory" (which can be thousands of times slower), or even **crash**. Or one process might need to do some computation and would have to wait until the CPU is free again.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sun Aug 31 09:15:41 UTC 2025 - 18.6K bytes - Viewed (0) -
docs/en/docs/deployment/server-workers.md
## Recap { #recap } You can use multiple worker processes with the `--workers` CLI option with the `fastapi` or `uvicorn` commands to take advantage of **multi-core CPUs**, to run **multiple processes in parallel**. You could use these tools and ideas if you are setting up **your own deployment system** while taking care of the other deployment concepts yourself.Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sun Aug 31 09:15:41 UTC 2025 - 8.3K bytes - Viewed (0) -
docs/ko/docs/deployment/docker.md
/// ### 공식 도커 이미지에 있는 프로세스 개수 이 이미지에 있는 **프로세스 개수**는 가용한 CPU **코어들**로 부터 **자동으로 계산**됩니다. 이것이 의미하는 바는 이미지가 CPU로부터 **최대한의 성능**을 **쥐어짜낸다**는 것입니다. 여러분은 이 설정 값을 **환경 변수**나 기타 방법들로 조정할 수 있습니다. 그러나 프로세스의 개수가 컨테이너가 실행되고 있는 CPU에 의존한다는 것은 또한 **소요되는 메모리의 크기** 또한 이에 의존한다는 것을 의미합니다.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Sat Nov 09 16:39:20 UTC 2024 - 42.7K bytes - Viewed (0) -
RELEASE.md
* `tf.raw_ops.Bucketize` op on CPU. * `tf.where` op for data types `tf.int32`/`tf.uint32`/`tf.int8`/`tf.uint8`/`tf.int64`. * `tf.random.normal` op for output data type `tf.float32` on CPU. * `tf.random.uniform` op for output data type `tf.float32` on CPU. * `tf.random.categorical` op for output data type `tf.int64` on CPU. * `tensorflow.experimental.tensorrt`:Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Mon Aug 18 20:54:38 UTC 2025 - 740K bytes - Viewed (2) -
docs/pt/docs/deployment/server-workers.md
## Recapitular Você pode usar vários processos de trabalho com a opção CLI `--workers` com os comandos `fastapi` ou `uvicorn` para aproveitar as vantagens de **CPUs multi-core** e executar **vários processos em paralelo**. Você pode usar essas ferramentas e ideias se estiver configurando **seu próprio sistema de implantação** enquanto cuida dos outros conceitos de implantação.
Registered: Sun Sep 07 07:19:17 UTC 2025 - Last Modified: Thu Jan 09 20:41:07 UTC 2025 - 8.5K bytes - Viewed (0) -
docs/metrics/prometheus/list.md
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Tue Aug 12 18:20:36 UTC 2025 - 43.4K bytes - Viewed (0) -
cmd/metrics-resource.go
Registered: Sun Sep 07 19:28:11 UTC 2025 - Last Modified: Sun Mar 30 00:56:02 UTC 2025 - 17.2K bytes - Viewed (0) -
tensorflow/c/README.md
- Nightly builds: - [Linux CPU-only](https://storage.googleapis.com/tensorflow-nightly/github/tensorflow/lib_package/libtensorflow-cpu-linux-x86_64.tar.gz) - [Linux GPU](https://storage.googleapis.com/tensorflow-nightly/github/tensorflow/lib_package/libtensorflow-gpu-linux-x86_64.tar.gz)
Registered: Tue Sep 09 12:39:10 UTC 2025 - Last Modified: Tue Oct 23 01:38:30 UTC 2018 - 539 bytes - Viewed (0)