Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 5 of 5 for Users (0.67 sec)

  1. RELEASE.md

          it's finished. The default is `False` for backward compatibility. Users of
          `distributed_save` are recommended to set it to `True`.
    
    * `tf.tpu.experimental.embedding.TPUEmbeddingV2`
        * Add `compute_sparse_core_stats` for sparse core users to profile the 
          data with this API to get the `max_ids` and `max_unique_ids`. These
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Apr 29 19:17:57 GMT 2024
    - 727.7K bytes
    - Viewed (8)
  2. tensorflow/c/BUILD

            ":tf_buffer_hdrs",
            ":tf_datatype_hdrs",
            ":tf_status_headers",
            ":tf_tensor_hdrs",
            # TODO: Only include tf_tstring_hdrs. Don't expose the implementation of TF_TString to API
            # users.
            ":tf_tstring",
            "//tensorflow/core:protos_all_cc",
        ],
    )
    
    tf_cuda_library(
        name = "c_api_internal",
        hdrs = [
            "c_api.h",
            "c_api_internal.h",
    Plain Text
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Wed Mar 27 18:00:18 GMT 2024
    - 30.3K bytes
    - Viewed (0)
  3. configure.py

      elif is_windows():
        default_cc_opt_flags = '/arch:AVX'
      else:
        # On all other platforms, no longer use `-march=native` as this can result
        # in instructions that are too modern being generated. Users that want
        # maximum performance should compile TF in their environment and can pass
        # `-march=native` there.
        # See https://github.com/tensorflow/tensorflow/issues/45744 and duplicates
    Python
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Mon Apr 15 18:25:36 GMT 2024
    - 53.8K bytes
    - Viewed (1)
  4. tensorflow/c/eager/tape.h

      // If `persistent` is true, GradientTape will not eagerly delete backward
      // functions (and hence the tensors they keep alive). Instead, everything
      // is deleted in ~GradientTape. Persistent GradientTapes are useful when
      // users want to compute multiple gradients over the same tape.
      explicit GradientTape(bool persistent) : persistent_(persistent) {}
      ~GradientTape() {
        for (const auto& pair : op_tape_) {
    C
    - Registered: Tue Apr 30 12:39:09 GMT 2024
    - Last Modified: Tue Apr 02 12:40:29 GMT 2024
    - 47.2K bytes
    - Viewed (1)
  5. .bazelrc

    # reduce the size of the debug info in the binary. This is because if the debug
    # sections in the ELF binary are too large, errors can occur. See
    # https://github.com/tensorflow/tensorflow/issues/48919.
    # Users can still include debug info for a specific kernel, e.g. with:
    #     --config=dbg --per_file_copt=+tensorflow/core/kernels/identity_op.*@-g
    # Since this .bazelrc file is synced between the tensorflow/tensorflow repo and
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Thu May 02 19:34:20 GMT 2024
    - 52.8K bytes
    - Viewed (2)
Back to top