- Sort Score
- Num 10 results
- Language All
Results 121 - 130 of 166 for cada (0.01 seconds)
-
WORKSPACE
load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl", "cuda_json_init_repository", ) cuda_json_init_repository() load( "@cuda_redist_json//:distributions.bzl", "CUDA_REDISTRIBUTIONS", "CUDNN_REDISTRIBUTIONS", ) load( "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Dec 26 23:20:26 GMT 2025 - 5.1K bytes - Click Count (0) -
.github/ISSUE_TEMPLATE/tflite-other.md
- type: input id: Compiler attributes: label: GCC/Compiler version description: if compiling from source placeholder: validations: required: false - type: input id: Cuda attributes: label: CUDA/cuDNN version description: placeholder: validations: required: false - type: input id: Gpu attributes: label: GPU model and memory description: ifCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Thu Dec 29 22:28:29 GMT 2022 - 3.4K bytes - Click Count (0) -
docs/es/docs/tutorial/path-params-numeric-validations.md
/// ## Ordena los parámetros como necesites { #order-the-parameters-as-you-need } /// tip | Consejo Esto probablemente no es tan importante o necesario si usas `Annotated`. /// Supongamos que quieres declarar el parámetro de query `q` como un `str` requerido.Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 20:41:43 GMT 2025 - 6.4K bytes - Click Count (0) -
tensorflow/BUILD
# Config setting that is satisfied when TensorFlow is being built with CUDA # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS). alias( name = "is_cuda_enabled", actual = if_oss( "@local_config_cuda//:is_cuda_enabled", "@local_config_cuda//cuda:using_config_cuda", ), ) # Config setting that is satisfied when CUDA device code should be compiledCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Wed Nov 12 19:21:56 GMT 2025 - 53.1K bytes - Click Count (0) -
ci/official/README.md
These "env" files match up with an environment matrix that roughly covers: - Different Python versions - Linux, MacOS, and Windows machines (these pool definitions are internal) - x86 and arm64 - CPU-only, or with NVIDIA CUDA support (Linux only), or with TPUs ## How to Test Your Changes to TensorFlow You may check how your changes will affect TensorFlow by: 1. Creating a PR and observing the presubmit test results
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Thu Feb 01 03:21:19 GMT 2024 - 8K bytes - Click Count (0) -
.github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml
description: If compiling from source - type: input id: Compiler attributes: label: GCC/compiler version description: If compiling from source - type: input id: Cuda attributes: label: CUDA/cuDNN version - type: input id: Gpu attributes: label: GPU model and memory description: If compiling from source - type: textarea id: what-happenedCreated: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Wed Jun 28 18:25:42 GMT 2023 - 3.7K bytes - Click Count (0) -
docs/pt/docs/tutorial/path-params-numeric-validations.md
/// note | Nota Um parâmetro de path é sempre obrigatório, pois precisa fazer parte do path. Mesmo que você o declare como `None` ou defina um valor padrão, isso não afetaria nada, ele ainda seria sempre obrigatório. /// ## Ordene os parâmetros de acordo com sua necessidade { #order-the-parameters-as-you-need } /// tip | DicaCreated: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 20:41:43 GMT 2025 - 6.7K bytes - Click Count (0) -
ci/official/requirements_updater/BUILD.bazel
load("@python_version_repo//:py_version.bzl", "REQUIREMENTS") load("@rules_python//python:pip.bzl", "compile_pip_requirements") # TODO(ybaturina): Remove once TF is migrated to CUDA 12.9. genrule( name = "nvidia_constraints", srcs = ["nvidia-requirements.txt"], outs = ["nvidia-constraints.txt"], cmd = """sed -E "s/>=/==/" $(location nvidia-requirements.txt) > $@;""", )Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Thu Sep 18 19:30:45 GMT 2025 - 1.4K bytes - Click Count (0) -
CONTRIBUTING.md
flag. ```bash export flags="--config=linux --config=cuda -k" ``` * For TensorFlow versions prior v.2.18.0: Add CUDA paths to LD_LIBRARY_PATH and add the `cuda` option flag. ```bash export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Sat Jan 11 04:47:59 GMT 2025 - 15.9K bytes - Click Count (0) -
ci/official/containers/ml_build/README.md
WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA). This container branches off from /tensorflow/tools/tf_sig_build_dockerfiles/. However, since hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the requirements installed on the original container can be removed to reduce the footprint of the container and make it more reusable across different ML
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Tue Sep 24 20:45:58 GMT 2024 - 416 bytes - Click Count (0)