Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 37 for bowker (0.16 sec)

  1. ci/official/utilities/cleanup_docker.sh

    cat <<EOF
    IMPORTANT: These tests ran under docker. This script does not clean up the
    container for you! You can delete the container with:
    
    $ docker rm -f tf
    
    You can also execute more commands within the container with e.g.:
    
    $ docker exec tf bazel clean
    $ docker exec -it tf bash
    EOF
    
    Shell Script
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Aug 10 20:26:29 GMT 2023
    - 998 bytes
    - Viewed (0)
  2. tensorflow/c/eager/parallel_device/parallel_device_remote_test.cc

      BasicTestsForTwoDevices(context.get(),
                              "/job:worker/replica:0/task:1/device:CPU:0",
                              "/job:worker/replica:0/task:2/device:CPU:0");
    
      worker_server1.release();
      worker_server2.release();
    }
    
    TEST(PARALLEL_DEVICE, TestAsyncCopyOff) {
      std::unique_ptr<TFE_ContextOptions, decltype(&TFE_DeleteContextOptions)> opts(
    C++
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Apr 27 22:09:57 GMT 2023
    - 6.7K bytes
    - Viewed (0)
  3. .github/workflows/arm-ci-extended-cpp.yml

            shell: bash
            run: |
              running_containers=$(docker ps -q) && \
              if [[ $running_containers == "" ]]; then
                echo "No running containers";
              else
                echo "Running container(s) found" && \
                docker stop $running_containers;
              fi
              docker container prune -f
              docker image prune -af
          - name: Clean repository
            shell: bash
    Others
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Wed Feb 07 17:41:21 GMT 2024
    - 2.5K bytes
    - Viewed (0)
  4. ci/official/envs/linux_arm64

    # despite lacking Nvidia CUDA support.
    TFCI_BUILD_PIP_PACKAGE_ARGS="--repo_env=WHEEL_NAME=tensorflow"
    TFCI_DOCKER_ENABLE=1
    TFCI_DOCKER_IMAGE=gcr.io/tensorflow-sigs/build-arm64:tf-2-16-multi-python
    TFCI_DOCKER_PULL_ENABLE=1
    TFCI_DOCKER_REBUILD_ARGS="--target=tf ci/official/containers/linux_arm64"
    TFCI_INDEX_HTML_ENABLE=1
    TFCI_LIB_SUFFIX="-cpu-linux-arm64"
    TFCI_OUTPUT_DIR=build_output
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Feb 15 23:12:40 GMT 2024
    - 1.5K bytes
    - Viewed (0)
  5. .github/workflows/arm-cd.yml

            shell: bash
            run: |
              running_containers=$(docker ps -q) && \
              if [[ $running_containers == "" ]]; then
                echo "No running containers";
              else
                echo "Running container(s) found" && \
                docker stop $running_containers;
              fi
              docker container prune -f
          - name: Clean repository
            shell: bash
    Others
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Tue Mar 05 10:24:16 GMT 2024
    - 3K bytes
    - Viewed (1)
  6. ci/official/containers/linux_arm64/cuda.packages.txt

    # CuDNN: https://docs.nvidia.com/deeplearning/sdk/cudnn-install/index.html#ubuntu-network-installation
    libcudnn8=8.9.6.50-1+cuda12.2
    libcudnn8-dev=8.9.6.50-1+cuda12.2
    
    # This can be removed once NVIDIA publishes a cuda-12.3.2 Docker image.
    # For now it ensures that we install at least version 12.3.107 of PTXAS,
    # since 12.3.103 has a bug.
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Mon Jan 08 09:32:19 GMT 2024
    - 368 bytes
    - Viewed (1)
  7. tensorflow/c/eager/immediate_execution_distributed_manager.h

                                          int64_t init_timeout_in_ms, int retries,
                                          bool clear_existing_contexts = false) = 0;
    
      // Initializes context for the local worker and no contexts will be created
      // for remote workers. Currently this only works for resetting context.
      // TODO(b/289445025): Consider removing this when we find a proper fix.
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Wed Feb 21 22:37:46 GMT 2024
    - 2.9K bytes
    - Viewed (0)
  8. ci/official/README.md

    #   container and start fresh, run "docker rm -f tf". Removing the container
    #   destroys some temporary bazel data and causes longer builds.
    #
    #   You will need the NVIDIA Container Toolkit for GPU testing:
    #   https://github.com/NVIDIA/nvidia-container-toolkit
    #
    #   Note: if you interrupt a bazel command on docker (ctrl-c), you
    #   will need to run `docker exec tf pkill bazel` to quit bazel.
    #
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Feb 01 03:21:19 GMT 2024
    - 8K bytes
    - Viewed (0)
  9. CONTRIBUTING.md

        execution time and the sharding
        [could create an overhead on the test execution](https://github.com/bazelbuild/bazel/issues/2113#issuecomment-264054799).
    
    2.  Using [Docker](https://www.docker.com) and TensorFlow's CI scripts.
    
        ```bash
        # Install Docker first, then this will build and run cpu tests
        tensorflow/tools/ci_build/ci_build.sh CPU bazel test //tensorflow/...
        ```
    
        See
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Mar 21 11:45:51 GMT 2024
    - 15.6K bytes
    - Viewed (0)
  10. ci/official/containers/linux_arm64/devel.usertools/aarch64_clang.bazelrc

    build --copt=-Wno-gnu-offsetof-extensions
    
    # Store performance profiling log in the mounted artifact directory.
    # The profile can be viewed by visiting chrome://tracing in a Chrome browser.
    # See https://docs.bazel.build/versions/main/skylark/performance.html#performance-profiling
    build --profile=/tf/pkg/profile.json.gz
    
    # Use the rebuilt gcc toolchain to compile for manylinux2014
    Plain Text
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Tue Nov 21 12:25:39 GMT 2023
    - 6.3K bytes
    - Viewed (0)
Back to top