Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 1,019 for Devices (0.11 sec)

  1. tensorflow/compiler/mlir/tensorflow/transforms/layout_optimization.cc

      func::FuncOp func = getOperation();
    
      // Get runtime devices information from the closest parent module.
      RuntimeDevices devices;
      if (failed(::tensorflow::GetDevicesFromOp(func->getParentOfType<ModuleOp>(),
                                                &devices)))
        return signalPassFailure();
    
      // If there is no runtime device information and data format is not explicitly
      // forced, there is nothing to do.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 19.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/tests/tpu_device_propagation.mlir

    // CHECK-SAME: ({{%.+}}: tensor<i64> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"}, {{%.+}}: tensor<i32> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"})
    // CHECK-SAME: -> (tensor<i64> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"}, tensor<i32> {tf.device = "/job:localhost/replica:0/task:0/device:TPU:0"})
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Mar 28 12:06:33 UTC 2022
    - 19K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/tests/replicate_invariant_op_hoisting.mlir

      %0:8 = tf_device.replicate([%arg0, %arg1] as %ri: tensor<*xf32>) {devices = {TPU_REPLICATED_CORE_0 = ["/device:TPU:0", "/device:TPU:1"]}, n = 2: i32} {
        %1 = "tf.Shape"(%ri) {device = "", T = "tfdtype$DT_FLOAT", out_type = "tfdtype$DT_INT32"} : (tensor<*xf32>) -> tensor<?xi32>
        %2 = "tf.opA"(%1) {device = "TPU_REPLICATED_CORE_0"} : (tensor<?xi32>) -> tensor<*xi32>
        %3 = "tf_device.launch"() ({
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:59:10 UTC 2023
    - 11.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_launch_util_gpu_test.cc

        // Set flag to use PJRT for device compilation and execution.
        auto& rollout_config = GetXlaOpsCommonFlags()->tf_xla_use_device_api;
        rollout_config.enabled_for_xla_launch_ = true;
        rollout_config.enabled_for_compile_on_demand_ = true;
        rollout_config.enabled_for_gpu_ = true;
    
        // Set flag to enable using XLA devices. PJRT currently is only supported
        // for XLA devices.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 10K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

        const auto sharding_type = sharding.type();
        if (sharding_type == xla::OpSharding::OTHER) {
          for (const auto& device : sharding.tile_assignment_devices()) {
            CHECK(device >= 0 && device < input_mappings.size());
            input_mappings[device].push_back(idx);
          }
        } else if (sharding_type == xla::OpSharding::REPLICATED) {
          for (auto& input : input_mappings) input.push_back(idx);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/transforms/cluster_tf_ops_pass.cc

    std::string GetHost(Operation *op) {
      std::string device = "";
      if (StringAttr attr = op->getAttrOfType<StringAttr>(kDeviceAttr)) {
        device = attr.getValue().str();
      }
      return GetHost(device);
    }
    
    // The device is considered to be on the localhost iff one of the following is
    // true:
    // 1) None of the job/replica/task is specified in the device name.
    // 2) The job/replica/task in the device name are explicitly specified as
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 13.4K bytes
    - Viewed (0)
  7. tensorflow/c/eager/parallel_device/parallel_device_test.cc

          TFE_ContextListDevices(context.get(), status.get()), TF_DeleteDeviceList);
      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
      bool has_tpu = false;
      for (int device_index = 0; device_index < TF_DeviceListCount(devices.get());
           ++device_index) {
        std::string device_type =
            TF_DeviceListType(devices.get(), device_index, status.get());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 08 23:47:35 UTC 2021
    - 29.3K bytes
    - Viewed (0)
  8. tensorflow/c/eager/c_api_experimental.cc

      // in an initialized context.
      for (auto d = devices.begin(); d != devices.end();) {
        if (absl::StrContains(d->get()->name(), "CPU:0")) {
          d = devices.erase(d);
        } else {
          ++d;
        }
      }
    
      status->status = tensorflow::unwrap(ctx)->AddDevices(std::move(devices));
    }
    
    void TFE_InsertConfigKeyValue(TFE_Context* ctx, const char* key,
                                  const char* value, TF_Status* status) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 11 23:52:39 UTC 2024
    - 35.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/partially_decluster_pass_test.cc

      // This is needed to register the XLA_GPU device.
      std::vector<std::unique_ptr<Device>> devices;
      TF_ASSERT_OK(DeviceFactory::AddDevices(
          SessionOptions(), "/job:localhost/replica:0/task:0", &devices));
    
      // Scope::ToGraph loses the assigned device name since it goes through
      // GraphDef/NodeDef which does not have a field for the assigned device name.
      Node* n = FindNodeByName(*graph, "shape");
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jun 10 12:32:39 UTC 2022
    - 23K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_collective.cc

                                    DenseIntElementsAttr replica_groups,
                                    Operation* op) {
      // Use special group_key 0 to represent "all available devices". This
      // shall resolve to a DeviceAssignment that includes all devices intended for
      // replica_groups.
      IntegerAttr group_size = builder.getI32IntegerAttr(replica_groups.size());
      IntegerAttr group_key = builder.getI32IntegerAttr(0);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 16K bytes
    - Viewed (0)
Back to top