Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 594 for Levine (0.17 sec)

  1. tensorflow/compiler/mlir/lite/experimental/tac/tests/device-transform-gpu.mlir

    // RUN: tac-opt-all-backends -tfl-device-transform-gpu %s -split-input-file -verify-diagnostics | FileCheck %s
    
    func.func @pack(%arg0: tensor<1xf32>, %arg1: tensor<1xf32>) -> tensor<2x1xf32> {
      %0 = "tfl.pack"(%arg0, %arg1) {axis = 0 : i32, values_count = 2 : i32} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32>
      func.return %0 : tensor<2x1xf32>
    }
    
    // CHECK:   func @pack(%[[VAL_0:.*]]: tensor<1xf32>, %[[VAL_1:.*]]: tensor<1xf32>) -> tensor<2x1xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 15.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/passes.h

    #define GEN_PASS_DECL_LAYOUTASSIGNMENTPASS
    #define GEN_PASS_DECL_LEGALIZEHLOTOTFPASS
    #define GEN_PASS_DECL_LEGALIZETFGTOTFPASS
    #define GEN_PASS_DECL_LOCALIZEVARHANDLESPASS
    #define GEN_PASS_DECL_LOWERQUANTIZEDPASS
    #define GEN_PASS_DECL_MARKINPUTOUTPUTALIASESPASS
    #define GEN_PASS_DECL_MATERIALIZEPASSTHROUGHOP
    #define GEN_PASS_DECL_MERGECONTROLFLOWPASS
    #define GEN_PASS_DECL_MOVETRANSPOSESPASS
    #define GEN_PASS_DECL_ORDERBYDIALECTPASS
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Jun 12 21:18:05 UTC 2024
    - 31.8K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.h

    limitations under the License.
    ==============================================================================*/
    
    #ifndef TENSORFLOW_COMPILER_MLIR_TF2XLA_API_V1_COMPILE_MLIR_UTIL_H_
    #define TENSORFLOW_COMPILER_MLIR_TF2XLA_API_V1_COMPILE_MLIR_UTIL_H_
    
    #include <memory>
    
    #include "absl/base/attributes.h"
    #include "llvm/ADT/ArrayRef.h"
    #include "llvm/ADT/StringRef.h"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 21 17:24:39 UTC 2024
    - 10.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/stablehlo/cc/saved_model_export_test.cc

              %c_9 = tf_executor.island wraps "tf.InitializeTableFromTextFileV2"(%o, %arg1) <{delimiter = "\09", key_index = -2 : i64, value_index = -1 : i64, vocab_size = -1 : i64}> {_has_manual_control_dependencies = true, device = ""} : (tensor<!tf_type.resource>, tensor<!tf_type.string>) -> ()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Mar 20 11:11:25 UTC 2024
    - 19.6K bytes
    - Viewed (0)
  5. tensorflow/c/kernels.cc

      if (cc_ctx->op_device_context() == nullptr) {  // CPU Device
        status->status = absl::FailedPreconditionError(
            "Accessing device stream is not supported for a CPU device.");
        return nullptr;
      } else if (!cc_ctx->op_device_context()->IsPluggableDevice()) {
        status->status = absl::FailedPreconditionError(
            "Accessing device stream is only supported for pluggable devices.");
        return nullptr;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 22:53:47 UTC 2024
    - 36K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h

    inline constexpr absl::string_view kDeviceAssignmentAttr = "device_assignment";
    
    // A TPU device for execution alongside its associated host CPU device.
    struct TPUDeviceAndHost {
      TPUDeviceAndHost() = default;
      TPUDeviceAndHost(llvm::StringRef device, llvm::StringRef host)
          : device(device), host(host) {}
    
      std::string device;
      std::string host;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 11.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/transforms/tpu_device_propagation.cc

        value_to_device.insert({result, device});
      }
    }
    
    // Checks if an operation can have TPU devices propagated through.
    bool IsSupportedOpToSetDevice(Operation& op) {
      return IsSupportedExecutorOp(op) ||
             isa<TF::IdentityOp, TF::IdentityNOp, TF::ShapeOp>(op);
    }
    
    // Finds nonconflicting TPU device for an operation from its operands. If an
    // operand has no device or a non TPU device, or if there are conflicting
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11K bytes
    - Viewed (0)
  8. tensorflow/c/eager/c_api_experimental.h

                                void* device_info) = nullptr;
    
      // Pins the op to `device` based on inputs to `op`. Returns true
      // signifying to pin to the current custom device. Returns false
      // to pin to the physical device.
      //
      // This function is guaranteed to be called only when all of the custom-device
      // inputs are on this device.
      bool (*shall_pin_to_this_device)(const TFE_Op* op, TF_Status* s) = nullptr;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 22:37:46 UTC 2024
    - 39.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/ir/tf_ops.cc

      auto node_device = [&](Operation *n) -> StringAttr {
        auto device = n->getAttrOfType<StringAttr>(device_id);
        if (!device || device.getValue().empty()) return caller_device;
    
        DeviceNameUtils::ParsedName ndef_parsed_device;
        if (!DeviceNameUtils::ParseFullName(device.getValue().str(),
                                            &ndef_parsed_device))
          return device;
        DeviceNameUtils::MergeUnsetDevNames(&ndef_parsed_device,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 14.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/xla_device.h

        // The name of the compilation device (e.g., "XLA_CPU_JIT");
        string compilation_device_name;
    
        // If 'use_multiple_streams' is true, we create separate streams for
        // compute, host-to-device, and device-to-host communication.
        bool use_multiple_streams = false;
    
        // If true, the XLA devices with the same device ordinal will share the same
        // compute stream. Otherwise each XLA device will having their own compute
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 13.4K bytes
    - Viewed (0)
Back to top