- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 1,019 for Devices (0.11 sec)
-
tensorflow/compiler/mlir/tensorflow/transforms/layout_optimization.cc
func::FuncOp func = getOperation(); // Get runtime devices information from the closest parent module. RuntimeDevices devices; if (failed(::tensorflow::GetDevicesFromOp(func->getParentOfType<ModuleOp>(), &devices))) return signalPassFailure(); // If there is no runtime device information and data format is not explicitly // forced, there is nothing to do.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 19.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/tpu_device_propagation.mlir
// CHECK-SAME: ({{%.+}}: tensor<i64> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"}, {{%.+}}: tensor<i32> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"}) // CHECK-SAME: -> (tensor<i64> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"}, tensor<i32> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"})
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon Mar 28 12:06:33 UTC 2022 - 19K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/replicate_invariant_op_hoisting.mlir
%0:8 = tf_device.replicate([%arg0, %arg1] as %ri: tensor<*xf32>) {devices = {TPU_REPLICATED_CORE_0 = ["/device:TPU:0", "/device:TPU:1"]}, n = 2: i32} { %1 = "tf.Shape"(%ri) {device = "", T = "tfdtype$DT_FLOAT", out_type = "tfdtype$DT_INT32"} : (tensor<*xf32>) -> tensor<?xi32> %2 = "tf.opA"(%1) {device = "TPU_REPLICATED_CORE_0"} : (tensor<?xi32>) -> tensor<*xi32> %3 = "tf_device.launch"() ({
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Oct 31 08:59:10 UTC 2023 - 11.9K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util_gpu_test.cc
// Set flag to use PJRT for device compilation and execution. auto& rollout_config = GetXlaOpsCommonFlags()->tf_xla_use_device_api; rollout_config.enabled_for_xla_launch_ = true; rollout_config.enabled_for_compile_on_demand_ = true; rollout_config.enabled_for_gpu_ = true; // Set flag to enable using XLA devices. PJRT currently is only supported // for XLA devices.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Sep 06 19:12:29 UTC 2023 - 10K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc
const auto sharding_type = sharding.type(); if (sharding_type == xla::OpSharding::OTHER) { for (const auto& device : sharding.tile_assignment_devices()) { CHECK(device >= 0 && device < input_mappings.size()); input_mappings[device].push_back(idx); } } else if (sharding_type == xla::OpSharding::REPLICATED) { for (auto& input : input_mappings) input.push_back(idx);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 21:28:13 UTC 2024 - 34K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/cluster_tf_ops_pass.cc
std::string GetHost(Operation *op) { std::string device = ""; if (StringAttr attr = op->getAttrOfType<StringAttr>(kDeviceAttr)) { device = attr.getValue().str(); } return GetHost(device); } // The device is considered to be on the localhost iff one of the following is // true: // 1) None of the job/replica/task is specified in the device name. // 2) The job/replica/task in the device name are explicitly specified as
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 13.4K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_test.cc
TFE_ContextListDevices(context.get(), status.get()), TF_DeleteDeviceList); ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get()); bool has_tpu = false; for (int device_index = 0; device_index < TF_DeviceListCount(devices.get()); ++device_index) { std::string device_type = TF_DeviceListType(devices.get(), device_index, status.get());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Jul 08 23:47:35 UTC 2021 - 29.3K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.cc
// in an initialized context. for (auto d = devices.begin(); d != devices.end();) { if (absl::StrContains(d->get()->name(), "CPU:0")) { d = devices.erase(d); } else { ++d; } } status->status = tensorflow::unwrap(ctx)->AddDevices(std::move(devices)); } void TFE_InsertConfigKeyValue(TFE_Context* ctx, const char* key, const char* value, TF_Status* status) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 11 23:52:39 UTC 2024 - 35.9K bytes - Viewed (0) -
tensorflow/compiler/jit/partially_decluster_pass_test.cc
// This is needed to register the XLA_GPU device. std::vector<std::unique_ptr<Device>> devices; TF_ASSERT_OK(DeviceFactory::AddDevices( SessionOptions(), "/job:localhost/replica:0/task:0", &devices)); // Scope::ToGraph loses the assigned device name since it goes through // GraphDef/NodeDef which does not have a field for the assigned device name. Node* n = FindNodeByName(*graph, "shape");
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jun 10 12:32:39 UTC 2022 - 23K bytes - Viewed (0) -
tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_collective.cc
DenseIntElementsAttr replica_groups, Operation* op) { // Use special group_key 0 to represent "all available devices". This // shall resolve to a DeviceAssignment that includes all devices intended for // replica_groups. IntegerAttr group_size = builder.getI32IntegerAttr(replica_groups.size()); IntegerAttr group_key = builder.getI32IntegerAttr(0);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 16K bytes - Viewed (0)