Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 14 of 14 for says (0.18 sec)

  1. .bazelrc

    # This is the same as the official TensorFlow builds.
    # See https://developer.nvidia.com/cuda-gpus#compute
    # `compute_XY` enables PTX embedding in addition to SASS. PTX
    # is forward compatible beyond the current compute capability major
    # release while SASS is only forward compatible inside the current
    # major release. Example: sm_80 kernels can run on sm_89 GPUs but
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Thu May 02 19:34:20 GMT 2024
    - 52.8K bytes
    - Viewed (2)
  2. ci/official/containers/linux_arm64/devel.usertools/squash_testlogs.py

    internal invocation viewer.
    """
    import collections
    import os
    import re
    import subprocess
    import sys
    from junitparser import JUnitXml
    
    result = JUnitXml()
    try:
      files = subprocess.check_output(
          ["grep", "-rlE", '(failures|errors)="[1-9]', sys.argv[1]]
      )
    except subprocess.CalledProcessError as e:
      print("No failures found to log!")
      exit(0)
    
    Python
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Sep 18 19:00:37 GMT 2023
    - 4.8K bytes
    - Viewed (0)
  3. tensorflow/c/experimental/filesystem/plugins/posix/copy_file.h

    ==============================================================================*/
    #ifndef TENSORFLOW_C_EXPERIMENTAL_FILESYSTEM_PLUGINS_POSIX_COPY_FILE_H_
    #define TENSORFLOW_C_EXPERIMENTAL_FILESYSTEM_PLUGINS_POSIX_COPY_FILE_H_
    
    #include <sys/stat.h>
    
    namespace tf_posix_filesystem {
    
    // Transfers up to `size` bytes from `dst_fd` to `src_fd`.
    //
    // This method uses `sendfile` if available (i.e., linux 2.6.33 or later) or an
    // intermediate buffer if not.
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Fri Nov 22 21:23:55 GMT 2019
    - 1.2K bytes
    - Viewed (0)
  4. tensorflow/c/eager/parallel_device/parallel_device.cc

        TF_Status* status) {
      absl::optional<std::vector<MaybeParallelTensorOwned>> result;
      // TODO(allenl): We should remove "TPU" from these op names at the very least,
      // or consider other ways of packing/unpacking parallel tensors.
      if (operation_name == std::string("TPUReplicatedInput")) {
        // Special-cased operation for packing per-device tensors into one parallel
        // tensor.
    C++
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 29 22:05:31 GMT 2023
    - 18.3K bytes
    - Viewed (0)
Back to top