- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 594 for Levine (0.17 sec)
-
tensorflow/compiler/mlir/lite/experimental/tac/tests/device-transform-gpu.mlir
// RUN: tac-opt-all-backends -tfl-device-transform-gpu %s -split-input-file -verify-diagnostics | FileCheck %s func.func @pack(%arg0: tensor<1xf32>, %arg1: tensor<1xf32>) -> tensor<2x1xf32> { %0 = "tfl.pack"(%arg0, %arg1) {axis = 0 : i32, values_count = 2 : i32} : (tensor<1xf32>, tensor<1xf32>) -> tensor<2x1xf32> func.return %0 : tensor<2x1xf32> } // CHECK: func @pack(%[[VAL_0:.*]]: tensor<1xf32>, %[[VAL_1:.*]]: tensor<1xf32>) -> tensor<2x1xf32> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 15.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/passes.h
#define GEN_PASS_DECL_LAYOUTASSIGNMENTPASS #define GEN_PASS_DECL_LEGALIZEHLOTOTFPASS #define GEN_PASS_DECL_LEGALIZETFGTOTFPASS #define GEN_PASS_DECL_LOCALIZEVARHANDLESPASS #define GEN_PASS_DECL_LOWERQUANTIZEDPASS #define GEN_PASS_DECL_MARKINPUTOUTPUTALIASESPASS #define GEN_PASS_DECL_MATERIALIZEPASSTHROUGHOP #define GEN_PASS_DECL_MERGECONTROLFLOWPASS #define GEN_PASS_DECL_MOVETRANSPOSESPASS #define GEN_PASS_DECL_ORDERBYDIALECTPASS
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 12 21:18:05 UTC 2024 - 31.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/api/v1/compile_mlir_util.h
limitations under the License. ==============================================================================*/ #ifndef TENSORFLOW_COMPILER_MLIR_TF2XLA_API_V1_COMPILE_MLIR_UTIL_H_ #define TENSORFLOW_COMPILER_MLIR_TF2XLA_API_V1_COMPILE_MLIR_UTIL_H_ #include <memory> #include "absl/base/attributes.h" #include "llvm/ADT/ArrayRef.h" #include "llvm/ADT/StringRef.h"
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 21 17:24:39 UTC 2024 - 10.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/quantization/stablehlo/cc/saved_model_export_test.cc
%c_9 = tf_executor.island wraps "tf.InitializeTableFromTextFileV2"(%o, %arg1) <{delimiter = "\09", key_index = -2 : i64, value_index = -1 : i64, vocab_size = -1 : i64}> {_has_manual_control_dependencies = true, device = ""} : (tensor<!tf_type.resource>, tensor<!tf_type.string>) -> ()
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Mar 20 11:11:25 UTC 2024 - 19.6K bytes - Viewed (0) -
tensorflow/c/kernels.cc
if (cc_ctx->op_device_context() == nullptr) { // CPU Device status->status = absl::FailedPreconditionError( "Accessing device stream is not supported for a CPU device."); return nullptr; } else if (!cc_ctx->op_device_context()->IsPluggableDevice()) { status->status = absl::FailedPreconditionError( "Accessing device stream is only supported for pluggable devices."); return nullptr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 22:53:47 UTC 2024 - 36K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h
inline constexpr absl::string_view kDeviceAssignmentAttr = "device_assignment"; // A TPU device for execution alongside its associated host CPU device. struct TPUDeviceAndHost { TPUDeviceAndHost() = default; TPUDeviceAndHost(llvm::StringRef device, llvm::StringRef host) : device(device), host(host) {} std::string device; std::string host; };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 26 09:37:10 UTC 2024 - 11.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_device_propagation.cc
value_to_device.insert({result, device}); } } // Checks if an operation can have TPU devices propagated through. bool IsSupportedOpToSetDevice(Operation& op) { return IsSupportedExecutorOp(op) || isa<TF::IdentityOp, TF::IdentityNOp, TF::ShapeOp>(op); } // Finds nonconflicting TPU device for an operation from its operands. If an // operand has no device or a non TPU device, or if there are conflicting
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
void* device_info) = nullptr; // Pins the op to `device` based on inputs to `op`. Returns true // signifying to pin to the current custom device. Returns false // to pin to the physical device. // // This function is guaranteed to be called only when all of the custom-device // inputs are on this device. bool (*shall_pin_to_this_device)(const TFE_Op* op, TF_Status* s) = nullptr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 22:37:46 UTC 2024 - 39.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.cc
auto node_device = [&](Operation *n) -> StringAttr { auto device = n->getAttrOfType<StringAttr>(device_id); if (!device || device.getValue().empty()) return caller_device; DeviceNameUtils::ParsedName ndef_parsed_device; if (!DeviceNameUtils::ParseFullName(device.getValue().str(), &ndef_parsed_device)) return device; DeviceNameUtils::MergeUnsetDevNames(&ndef_parsed_device,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.6K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.h
// The name of the compilation device (e.g., "XLA_CPU_JIT"); string compilation_device_name; // If 'use_multiple_streams' is true, we create separate streams for // compute, host-to-device, and device-to-host communication. bool use_multiple_streams = false; // If true, the XLA devices with the same device ordinal will share the same // compute stream. Otherwise each XLA device will having their own compute
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 13.4K bytes - Viewed (0)