Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 26 for Suda (0.16 sec)

  1. ci/official/envs/linux_arm64

    TFCI_BAZEL_TARGET_SELECTING_CONFIG_PREFIX=linux_arm64
    # Note: this is not set to "--cpu", because that changes the package name
    # to tensorflow_cpu. These ARM builds are supposed to have the name "tensorflow"
    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=gcr.io/tensorflow-sigs/build-arm64:tf-2-16-multi-python
    TFCI_DOCKER_PULL_ENABLE=1
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 23:12:40 GMT 2024
    - 1.5K bytes
    - Viewed (1)
  2. CONTRIBUTING.md

        export flags="--config=opt -k"
        ```
    
        If the tests are to be run on the GPU, add CUDA paths to LD_LIBRARY_PATH and
        add the `cuda` option flag
    
        ```bash
        export LD_LIBRARY_PATH="${LD_LIBRARY_PATH}:/usr/local/cuda/lib64:/usr/local/cuda/extras/CUPTI/lib64:$LD_LIBRARY_PATH"
        export flags="--config=opt --config=cuda -k"
        ```
    
        For example, to run all tests under tensorflow/python, do:
    
        ```bash
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Thu Mar 21 11:45:51 GMT 2024
    - 15.6K bytes
    - Viewed (0)
  3. ci/official/containers/linux_arm64/devel.usertools/aarch64_clang.bazelrc

    # Change the value of CACHEBUSTER when upgrading the toolchain, or when testing
    # different compilation methods. E.g. for a PR to test a new CUDA version, set
    # the CACHEBUSTER to the PR number.
    build --action_env=CACHEBUSTER=20220325
    
    # Use Python 3.X as installed in container image
    build --action_env PYTHON_BIN_PATH="/usr/local/bin/python3"
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Tue Nov 21 12:25:39 GMT 2023
    - 6.3K bytes
    - Viewed (0)
  4. ci/official/utilities/rename_and_verify_wheels.sh

    fi
    # VERY basic check to ensure the [and-cuda] package variant is installable.
    # Checks TFCI_BAZEL_COMMON_ARGS for "gpu" or "cuda", implying that the test is
    # relevant. All of the GPU test machines have CUDA installed via other means,
    # so I am not sure how to verify that the dependencies themselves are valid for
    # the moment.
    if [[ "$TFCI_BAZEL_COMMON_ARGS" =~ gpu|cuda ]]; then
    Shell Script
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 27 21:16:27 GMT 2024
    - 3.4K bytes
    - Viewed (0)
  5. tensorflow/BUILD

    # Config setting that is satisfied when TensorFlow is being built with CUDA
    # support through e.g. `--config=cuda` (or `--config=cuda_clang` in OSS).
    alias(
        name = "is_cuda_enabled",
        actual = if_oss(
            "@local_config_cuda//:is_cuda_enabled",
            "@local_config_cuda//cuda:using_clang",
        ),
    )
    
    # Config setting that is satisfied when CUDA device code should be compiled
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Apr 09 18:15:11 GMT 2024
    - 53.4K bytes
    - Viewed (8)
  6. tensorflow/c/eager/BUILD

    load(
        "//tensorflow:tensorflow.bzl",
        "tf_cc_test",
        "tf_copts",
        "tf_cuda_cc_test",
        "tf_cuda_library",
    )
    load("//tensorflow:tensorflow.default.bzl", "cc_header_only_library", "filegroup")
    load(
        "//tensorflow/core/platform:build_config_root.bzl",
        "tf_cuda_tests_tags",
    )
    load("//tensorflow/core/platform:rules_cc.bzl", "cc_library")
    
    package(
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Thu Apr 11 23:52:39 GMT 2024
    - 33.3K bytes
    - Viewed (0)
  7. ci/official/libtensorflow.sh

    # limitations under the License.
    # ==============================================================================
    source "${BASH_SOURCE%/*}/utilities/setup.sh"
    
    # Record GPU count and CUDA version status
    if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then
      tfrun nvidia-smi
    fi
    
    # Update the version numbers for Nightly only
    if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
    Shell Script
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Fri Jan 19 19:07:48 GMT 2024
    - 1.5K bytes
    - Viewed (0)
  8. ci/official/wheel.sh

    # limitations under the License.
    # ==============================================================================
    source "${BASH_SOURCE%/*}/utilities/setup.sh"
    
    # Record GPU count and CUDA version status
    if [[ "$TFCI_NVIDIA_SMI_ENABLE" == 1 ]]; then
      tfrun nvidia-smi
    fi
    
    # Update the version numbers for Nightly only
    if [[ "$TFCI_NIGHTLY_UPDATE_VERSION_ENABLE" == 1 ]]; then
    Shell Script
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 06 21:54:13 GMT 2024
    - 1.8K bytes
    - Viewed (0)
  9. RELEASE.md

    *   Move `layers_dense_variational_impl.py` to `layers_dense_variational.py`.
    
    ## Known Bugs
    
    *   Using XLA:GPU with CUDA 9 and CUDA 9.1 results in garbage results and/or
        `CUDA_ILLEGAL_ADDRESS` failures.
    
        Google discovered in mid-December 2017 that the PTX-to-SASS compiler in CUDA
        9 and CUDA 9.1 sometimes does not properly compute the carry bit when
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Apr 29 19:17:57 GMT 2024
    - 727.7K bytes
    - Viewed (8)
  10. configure.py

    
    def set_tf_cuda_version(environ_cp):
      """Set TF_CUDA_VERSION."""
      ask_cuda_version = (
          'Please specify the CUDA SDK version you want to use. '
          '[Leave empty to default to CUDA %s]: ') % _DEFAULT_CUDA_VERSION
      tf_cuda_version = get_from_env_or_user_or_default(environ_cp,
                                                        'TF_CUDA_VERSION',
    Python
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Apr 15 18:25:36 GMT 2024
    - 53.8K bytes
    - Viewed (1)
Back to top