- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 1,105 for Levine (0.12 sec)
-
tensorflow/c/kernels.cc
if (cc_ctx->op_device_context() == nullptr) { // CPU Device status->status = absl::FailedPreconditionError( "Accessing device stream is not supported for a CPU device."); return nullptr; } else if (!cc_ctx->op_device_context()->IsPluggableDevice()) { status->status = absl::FailedPreconditionError( "Accessing device stream is only supported for pluggable devices."); return nullptr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 22:53:47 UTC 2024 - 36K bytes - Viewed (0) -
tensorflow/cc/experimental/libtf/tests/function_test.cc
FunctionDef SquareFunc() { return FunctionDefHelper::Define( // Function Name "SquareFunc", // Args {"x: float"}, // Returns {"y: float"}, // Attr def {}, // Nodes {{/*ret=*/{"y"}, /*op=*/"Square", /*arg=*/{"x"}, /*attr=*/{{"T", DT_FLOAT}}, /*dep=*/{}, /*device=*/"", /*name=*/"square"}}); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 19 21:44:52 UTC 2023 - 11.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h
inline constexpr absl::string_view kDeviceAssignmentAttr = "device_assignment"; // A TPU device for execution alongside its associated host CPU device. struct TPUDeviceAndHost { TPUDeviceAndHost() = default; TPUDeviceAndHost(llvm::StringRef device, llvm::StringRef host) : device(device), host(host) {} std::string device; std::string host; };
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 26 09:37:10 UTC 2024 - 11.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_device_propagation.cc
value_to_device.insert({result, device}); } } // Checks if an operation can have TPU devices propagated through. bool IsSupportedOpToSetDevice(Operation& op) { return IsSupportedExecutorOp(op) || isa<TF::IdentityOp, TF::IdentityNOp, TF::ShapeOp>(op); } // Finds nonconflicting TPU device for an operation from its operands. If an // operand has no device or a non TPU device, or if there are conflicting
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
void* device_info) = nullptr; // Pins the op to `device` based on inputs to `op`. Returns true // signifying to pin to the current custom device. Returns false // to pin to the physical device. // // This function is guaranteed to be called only when all of the custom-device // inputs are on this device. bool (*shall_pin_to_this_device)(const TFE_Op* op, TF_Status* s) = nullptr;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 22:37:46 UTC 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_context.h
virtual void SetExecutorForThread(EagerExecutor* executor) = 0; // Return a list of local tensorflow::Device*. // TODO(tfrt-devs): We shouldn't expose legacy device in this API. virtual std::vector<tensorflow::Device*> ListLocalTfDevices() = 0; // Return a list of all tensorflow::Device*. virtual std::vector<tensorflow::Device*> ListAllTfDevices() = 0; //===--------------------------------------------------------------------===//
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 06 08:34:00 UTC 2023 - 12.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_ops.cc
auto node_device = [&](Operation *n) -> StringAttr { auto device = n->getAttrOfType<StringAttr>(device_id); if (!device || device.getValue().empty()) return caller_device; DeviceNameUtils::ParsedName ndef_parsed_device; if (!DeviceNameUtils::ParseFullName(device.getValue().str(), &ndef_parsed_device)) return device; DeviceNameUtils::MergeUnsetDevNames(&ndef_parsed_device,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 14.6K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.h
// The name of the compilation device (e.g., "XLA_CPU_JIT"); string compilation_device_name; // If 'use_multiple_streams' is true, we create separate streams for // compute, host-to-device, and device-to-host communication. bool use_multiple_streams = false; // If true, the XLA devices with the same device ordinal will share the same // compute stream. Otherwise each XLA device will having their own compute
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 13.4K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_device_ops.td
{n = 2 : i32, devices = {DEVICE_ALIAS_0 = ["/DEVICE:0", "/DEVICE:1"], DEVICE_ALIAS_1 = ["/DEVICE:2", "/DEVICE:3"]}} { // Inside the region, %0, %2, %4, and %6 corresponds to // "/DEVICE:0"/"/DEVICE:2" and %1, %3, %5, and %7 corresponds to // "/DEVICE:1"/"/DEVICE:3", depending on which device alias is used. %k = "tf_device.launch"() ( {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jan 23 23:53:20 UTC 2024 - 14.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/ir/tfrt_fallback_async.td
args: a list of fallback tensors. device: the name of the tensorflow device that is associated with the input fallback tensors. results: the converted corert tensorhandles. }]; let arguments = (ins Variadic<TFTensorType>:$args, StrAttr:$device ); let results = (outs Variadic<CoreRT_TensorHandleType>:$results );
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 18 15:01:21 UTC 2024 - 15.8K bytes - Viewed (0)