Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 18 for Suda (0.2 sec)

  1. ci/official/containers/linux_arm64/cuda.packages.txt

    libcudnn8=8.9.6.50-1+cuda12.2
    libcudnn8-dev=8.9.6.50-1+cuda12.2
    
    # This can be removed once NVIDIA publishes a cuda-12.3.2 Docker image.
    # For now it ensures that we install at least version 12.3.107 of PTXAS,
    # since 12.3.103 has a bug.
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Jan 08 09:32:19 GMT 2024
    - 368 bytes
    - Viewed (1)
  2. ci/official/containers/linux_arm64/Dockerfile

    # Install devtoolset devel dependencies
    COPY setup.sources.sh /setup.sources.sh
    COPY setup.packages.sh /setup.packages.sh
    COPY devel.packages.txt /devel.packages.txt
    COPY cuda.packages.txt /cuda.packages.txt
    RUN /setup.sources.sh && /setup.packages.sh /devel.packages.txt
    
    # Install various tools.
    # - bats: bash unit testing framework
    #         NOTE: v1.6.0 seems to have a bug that made "git" in setup_file break
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Jan 08 09:32:19 GMT 2024
    - 4.1K bytes
    - Viewed (1)
  3. ci/official/envs/linux_arm64

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64
    # Note: this is not set to "--cpu", because that changes the package name
    # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow"
    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=gcr.io/tensorflow-sigs/build-arm64:tf-2-16-multi-python
    TFCI_DOCKER_PULL_ENABLE=1
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 23:12:40 GMT 2024
    - 1.5K bytes
    - Viewed (1)
  4. ci/official/containers/linux_arm64/devel.usertools/aarch64_clang.bazelrc

    # Change the value of CACHEBUSTER when upgrading the toolchain, or when testing
    # different compilation methods. E.g. for a PR to test a new CUDA version, set
    # the CACHEBUSTER to the PR number.
    build --action_env=CACHEBUSTER=20220325
    
    # Use Python 3.X as installed in container image
    build --action_env PYTHON_BIN_PATH="/usr/local/bin/python3"
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Nov 21 12:25:39 GMT 2023
    - 6.3K bytes
    - Viewed (0)
  5. ci/official/envs/linux_x86_cuda

    source ci/official/envs/linux_x86
    TFCI_BAZEL_COMMON_ARGS="--repo_env=TF_PYTHON_VERSION=$TFCI_PYTHON_VERSION --config release_gpu_linux"
    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_cuda
    TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ARGS="--gpus all"
    TFCI_LIB_SUFFIX="-gpu-linux-x86_64"
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Jan 19 00:24:30 GMT 2024
    - 1K bytes
    - Viewed (0)
  6. ci/official/containers/linux_arm64/devel.usertools/aarch64.bazelrc

    # Change the value of CACHEBUSTER when upgrading the toolchain, or when testing
    # different compilation methods. E.g. for a PR to test a new CUDA version, set
    # the CACHEBUSTER to the PR number.
    build --action_env=CACHEBUSTER=20220325
    
    # Use Python 3.X as installed in container image
    build --action_env PYTHON_BIN_PATH="/usr/local/bin/python3"
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Nov 21 12:25:39 GMT 2023
    - 5.8K bytes
    - Viewed (0)
  7. docs/pt/docs/python-types.md

    ```Python
        first_name="john", last_name="doe"
    ```
    
    É uma coisa diferente.
    
    Estamos usando dois pontos (`:`), não é igual a (`=`).
    
    E adicionar type hints normalmente não muda o que acontece do que aconteceria sem elas.
    
    Mas agora, imagine que você está novamente no meio da criação dessa função, mas com type hints.
    
    No mesmo ponto, você tenta acionar o preenchimento automático com o `Ctrl Space` e vê:
    Plain Text
    - Registered: Sun Apr 28 07:19:10 GMT 2024
    - Last Modified: Fri Mar 22 01:42:11 GMT 2024
    - 9.6K bytes
    - Viewed (0)
  8. okhttp/src/test/resources/okhttp3/internal/publicsuffix/public_suffix_list.dat

    kurotaki.nara.jp
    mitsue.nara.jp
    miyake.nara.jp
    nara.nara.jp
    nosegawa.nara.jp
    oji.nara.jp
    ouda.nara.jp
    oyodo.nara.jp
    sakurai.nara.jp
    sango.nara.jp
    shimoichi.nara.jp
    shimokitayama.nara.jp
    shinjo.nara.jp
    soni.nara.jp
    takatori.nara.jp
    tawaramoto.nara.jp
    tenkawa.nara.jp
    tenri.nara.jp
    uda.nara.jp
    yamatokoriyama.nara.jp
    yamatotakada.nara.jp
    yamazoe.nara.jp
    yoshino.nara.jp
    Plain Text
    - Registered: Fri May 03 11:42:14 GMT 2024
    - Last Modified: Wed Dec 20 23:27:07 GMT 2023
    - 240.3K bytes
    - Viewed (3)
  9. tensorflow/c/BUILD

    load(
        "//tensorflow:tensorflow.bzl",
        "check_deps",
        "if_google",
        "if_not_mobile",
        "tf_cc_test",
        "tf_copts",
        "tf_cuda_library",
        "tf_custom_op_library",
        "tf_kernel_library",
    )
    load("//tensorflow:tensorflow.default.bzl", "filegroup", "tf_cuda_cc_test")
    load(
        "//tensorflow/core/tpu:build_defs.bzl",
        "if_libtpu_tf_status",
    )
    
    package(
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 27 18:00:18 GMT 2024
    - 30.3K bytes
    - Viewed (0)
  10. ci/official/README.md

    #    build. This should also match the system you're using--you cannot build
    #    the TF MacOS package from Linux.
    #      Ex. linux_x86        -- x86_64 Linux platform
    #      Ex. linux_x86_cuda   -- x86_64 Linux platform, with Nvidia CUDA support
    #      Ex. macos_arm64      -- arm64 MacOS platform
    # 3. Add modifiers. Some modifiers for local execution are:
    #      Ex. disk_cache -- Use a local cache
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 01 03:21:19 GMT 2024
    - 8K bytes
    - Viewed (0)
Back to top