Search Options

Display Count
Sort
Preferred Language
Advanced Search

Results 121 - 130 of 166 for cada (0.03 seconds)

  1. WORKSPACE

    load(
        "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_json_init_repository.bzl",
        "cuda_json_init_repository",
    )
    
    cuda_json_init_repository()
    
    load(
        "@cuda_redist_json//:distributions.bzl",
        "CUDA_REDISTRIBUTIONS",
        "CUDNN_REDISTRIBUTIONS",
    )
    load(
        "@rules_ml_toolchain//third_party/gpus/cuda/hermetic:cuda_redist_init_repositories.bzl",
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Fri Dec 26 23:20:26 GMT 2025
    - 5.1K bytes
    - Click Count (0)
  2. .github/ISSUE_TEMPLATE/tflite-other.md

    -   type: input id: Compiler attributes: label: GCC/Compiler version
        description: if compiling from source placeholder: validations: required:
        false
    
    -   type: input id: Cuda attributes: label: CUDA/cuDNN version description:
        placeholder: validations: required: false
    
    -   type: input id: Gpu attributes: label: GPU model and memory description: if
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Thu Dec 29 22:28:29 GMT 2022
    - 3.4K bytes
    - Click Count (0)
  3. docs/es/docs/tutorial/path-params-numeric-validations.md

    ///
    
    ## Ordena los parámetros como necesites { #order-the-parameters-as-you-need }
    
    /// tip | Consejo
    
    Esto probablemente no es tan importante o necesario si usas `Annotated`.
    
    ///
    
    Supongamos que quieres declarar el parámetro de query `q` como un `str` requerido.
    
    Created: Sun Dec 28 07:19:09 GMT 2025
    - Last Modified: Wed Dec 17 20:41:43 GMT 2025
    - 6.4K bytes
    - Click Count (0)
  4. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_config_cuda",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Wed Nov 12 19:21:56 GMT 2025
    - 53.1K bytes
    - Click Count (0)
  5. ci/official/README.md

    These "env" files match up with an environment matrix that roughly covers:
    
    -   Different Python versions
    -   Linux, MacOS, and Windows machines (these pool definitions are internal)
    -   x86 and arm64
    -   CPU-only, or with NVIDIA CUDA support (Linux only), or with TPUs
    
    ## How to Test Your Changes to TensorFlow
    
    You may check how your changes will affect TensorFlow by:
    
    1. Creating a PR and observing the presubmit test results
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Thu Feb 01 03:21:19 GMT 2024
    - 8K bytes
    - Click Count (0)
  6. .github/ISSUE_TEMPLATE/tensorflow_issue_template.yaml

          description: If compiling from source
      - type: input
        id: Compiler
        attributes:
          label: GCC/compiler version
          description: If compiling from source
      - type: input
        id: Cuda
        attributes:
          label: CUDA/cuDNN version
      - type: input
        id: Gpu
        attributes:
          label: GPU model and memory
          description: If compiling from source
      - type: textarea
        id: what-happened
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Wed Jun 28 18:25:42 GMT 2023
    - 3.7K bytes
    - Click Count (0)
  7. docs/pt/docs/tutorial/path-params-numeric-validations.md

    /// note | Nota
    
    Um parâmetro de path é sempre obrigatório, pois precisa fazer parte do path. Mesmo que você o declare como `None` ou defina um valor padrão, isso não afetaria nada, ele ainda seria sempre obrigatório.
    
    ///
    
    ## Ordene os parâmetros de acordo com sua necessidade { #order-the-parameters-as-you-need }
    
    /// tip | Dica
    
    Created: Sun Dec 28 07:19:09 GMT 2025
    - Last Modified: Wed Dec 17 20:41:43 GMT 2025
    - 6.7K bytes
    - Click Count (0)
  8. ci/official/requirements_updater/BUILD.bazel

    load("@python_version_repo//:py_version.bzl", "REQUIREMENTS")
    load("@rules_python//python:pip.bzl", "compile_pip_requirements")
    
    # TODO(ybaturina): Remove once TF is migrated to CUDA 12.9.
    genrule(
        name = "nvidia_constraints",
        srcs = ["nvidia-requirements.txt"],
        outs = ["nvidia-constraints.txt"],
        cmd = """sed -E "s/>=/==/" $(location nvidia-requirements.txt) > $@;""",
    )
    
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Thu Sep 18 19:30:45 GMT 2025
    - 1.4K bytes
    - Click Count (0)
  9. CONTRIBUTING.md

            flag.
    
            ```bash
            export flags="--config=linux --config=cuda -k"
            ```
    
        *   For TensorFlow versions prior v.2.18.0: Add CUDA paths to
            LD_LIBRARY_PATH and add the `cuda` option flag.
    
            ```bash
            export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Sat Jan 11 04:47:59 GMT 2025
    - 15.9K bytes
    - Click Count (0)
  10. ci/official/containers/ml_build/README.md

    WIP ML Build Docker container for ML repositories (Tensorflow, JAX and XLA).
    
    This container branches off from
    /tensorflow/tools/tf_sig_build_dockerfiles/. However, since
    hermetic CUDA and hermetic Python is now available for Tensorflow, a lot of the
    requirements installed on the original container can be removed to reduce the
    footprint of the container and make it more reusable across different ML
    Created: Tue Dec 30 12:39:10 GMT 2025
    - Last Modified: Tue Sep 24 20:45:58 GMT 2024
    - 416 bytes
    - Click Count (0)
Back to Top