- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 51 for cpus (0.02 seconds)
-
docs/es/docs/deployment/server-workers.md
## Resumen { #recap } Puedes usar múltiples worker processes con la opción CLI `--workers` con los comandos `fastapi` o `uvicorn` para aprovechar los **CPUs de múltiples núcleos**, para ejecutar **múltiples procesos en paralelo**.Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Tue Dec 16 16:33:45 GMT 2025 - 8.5K bytes - Click Count (0) -
docs/de/docs/deployment/server-workers.md
## Zusammenfassung { #recap } Sie können mehrere Workerprozesse mit der `--workers`-CLI-Option über die `fastapi`- oder `uvicorn`-Befehle nutzen, um **Multikern-CPUs** auszunutzen und **mehrere Prozesse parallel** auszuführen. Sie könnten diese Tools und Ideen nutzen, wenn Sie **Ihr eigenes Deployment-System** einrichten und sich dabei selbst um die anderen Deployment-Konzepte kümmern.Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Tue Dec 02 17:32:56 GMT 2025 - 8.8K bytes - Click Count (0) -
docs/es/docs/deployment/concepts.md
Por otro lado, si tienes 2 servidores y estás usando **100% de su CPU y RAM**, en algún momento un proceso pedirá más memoria y el servidor tendrá que usar el disco como "memoria" (lo cual puede ser miles de veces más lento), o incluso **colapsar**. O un proceso podría necesitar hacer algún cálculo y tendría que esperar hasta que la CPU esté libre de nuevo.
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Tue Dec 16 16:33:45 GMT 2025 - 20.1K bytes - Click Count (0) -
docs/de/docs/deployment/concepts.md
Wie viele Systemressourcen möchten Sie verbrauchen/nutzen? Sie mögen „nicht viel“ denken, aber in Wirklichkeit möchten Sie tatsächlich **so viel wie möglich ohne Absturz** verwenden.
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Tue Dec 02 17:32:56 GMT 2025 - 21.5K bytes - Click Count (0) -
.bazelrc
# See https://developer.nvidia.com/cuda-gpus#compute # `compute_XY` enables PTX embedding in addition to SASS. PTX # is forward compatible beyond the current compute capability major # release while SASS is only forward compatible inside the current # major release. Example: sm_80 kernels can run on sm_89 GPUs but # not on sm_90 GPUs. compute_80 kernels though can also run on sm_90 GPUs.
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 56.8K bytes - Click Count (0) -
WORKSPACE
load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl", "cuda_json_init_repository", ) cuda_json_init_repository() load( "@cuda_redist_json//:distributions.bzl", "CUDA_REDISTRIBUTIONS", "CUDNN_REDISTRIBUTIONS", ) load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 5.1K bytes - Click Count (0) -
ci/official/envs/linux_arm64
TFCI_BAZEL_COMMON_ARGS="--repo_env=HERMETIC_PYTHON_VERSION=$TFCI_PYTHON_VERSION --repo_env=USE_PYWRAP_RULES=True --config release_arm64_linux" TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64 # Note: this is not set to "--cpu", because that changes the package name # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow" # despite lacking Nvidia CUDA support. TFCI_BUILD_PIP_PACKAGE_WHEEL_NAME_ARG="--repo_env=WHEEL_NAME=tensorflow"
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Dec 13 00:14:04 GMT 2025 - 1.6K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/helper/SystemHelper.java
} } /** * Calibrates the CPU load. * * @return true if the CPU load is within the acceptable range, false otherwise. */ public boolean calibrateCpuLoad() { return calibrateCpuLoad(0L); } /** * Calibrates the CPU load with a timeout. * * @param timeoutInMillis The timeout in milliseconds.Created: Sat Dec 20 09:19:18 GMT 2025 - Last Modified: Sat Dec 20 08:30:43 GMT 2025 - 36.6K bytes - Click Count (0) -
.github/workflows/arm-ci.yml
run: | CI_DOCKER_BUILD_EXTRA_PARAMS="--pull --build-arg py_major_minor_version=${{ matrix.pyver }} --build-arg is_nightly=1 --build-arg tf_project_name=tf_nightly_ci" \Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Dec 01 09:57:00 GMT 2025 - 2.2K bytes - Click Count (0) -
.github/workflows/arm-ci-extended.yml
CI_DOCKER_BUILD_EXTRA_PARAMS="--build-arg py_major_minor_version=${{ matrix.pyver }} --build-arg is_nightly=${is_nightly} --build-arg tf_project_name=${tf_project_name}" \Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Dec 01 09:57:00 GMT 2025 - 2.6K bytes - Click Count (0)