Search Options

Results per page
Sort
Preferred Languages
Advance

Results 41 - 50 of 155 for r2devices (0.19 sec)

  1. tensorflow/c/eager/parallel_device/parallel_device_lib.cc

      }
    }
    
    ParallelDevice::ParallelDevice(const std::vector<std::string>& devices,
                                   bool is_async, int in_flight_nodes_limit)
        : underlying_devices_(devices),
          default_cancellation_manager_(absl::make_unique<CancellationManager>()) {
      device_threads_.reserve(devices.size());
      for (int device_index = 0; device_index < devices.size(); ++device_index) {
        device_threads_.emplace_back(new DeviceThread(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Feb 09 07:47:20 UTC 2024
    - 25.4K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/layout_optimization.cc

    void LayoutAssignmentPass::runOnOperation() {
      func::FuncOp func = getOperation();
    
      // Get runtime devices information from the closest parent module.
      RuntimeDevices devices;
      if (failed(::tensorflow::GetDevicesFromOp(func->getParentOfType<ModuleOp>(),
                                                &devices)))
        return signalPassFailure();
    
      // If there is no runtime device information and data format is not explicitly
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 19.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc

          location, concat_dim_type, concat_dimension_attr);
    
      // Correctly set output shapes of concat op output if output shape is
      // statically known. Since the shape of TPUExecute op must be the same
      // across logical devices, we refer to the shape of 0th logical device
      // computation output.
      mlir::Type output_type;
      auto input_type = mlir::cast<mlir::TensorType>(inputs[0].getType());
    
      if (input_type.hasRank()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 22 21:28:13 UTC 2024
    - 34K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/tests/replicate_invariant_op_hoisting.mlir

    func.func @do_not_hoist_ops_with_virtual_device(%arg0: tensor<*xf32>, %arg1: tensor<*xf32>) {
      %0:8 = tf_device.replicate([%arg0, %arg1] as %ri: tensor<*xf32>) {devices = {TPU_REPLICATED_CORE_0 = ["/device:TPU:0", "/device:TPU:1"]}, n = 2: i32} {
        %1 = "tf.Shape"(%ri) {device = "", T = "tfdtype$DT_FLOAT", out_type = "tfdtype$DT_INT32"} : (tensor<*xf32>) -> tensor<?xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:59:10 UTC 2023
    - 11.9K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/xla_launch_util_gpu_test.cc

        rollout_config.enabled_for_xla_launch_ = true;
        rollout_config.enabled_for_compile_on_demand_ = true;
        rollout_config.enabled_for_gpu_ = true;
    
        // Set flag to enable using XLA devices. PJRT currently is only supported
        // for XLA devices.
        GetXlaDeviceFlags()->tf_xla_enable_xla_devices = true;
    
        // Add and setup the GPU device.
        auto device_type = DeviceType(DEVICE_GPU);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 10K bytes
    - Viewed (0)
  6. tensorflow/c/eager/c_api_experimental.cc

      // in an initialized context.
      for (auto d = devices.begin(); d != devices.end();) {
        if (absl::StrContains(d->get()->name(), "CPU:0")) {
          d = devices.erase(d);
        } else {
          ++d;
        }
      }
    
      status->status = tensorflow::unwrap(ctx)->AddDevices(std::move(devices));
    }
    
    void TFE_InsertConfigKeyValue(TFE_Context* ctx, const char* key,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 11 23:52:39 UTC 2024
    - 35.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tensorflow/transforms/cluster_tf_ops_pass.cc

      llvm::SmallVector<Value, 4> inputs;
      // The result values of the function.
      llvm::SmallVector<Value, 4> results;
      // The devices of the input values. It should have the same size as inputs.
      llvm::SmallVector<std::string, 4> input_devices;
      // The devices of the result values. It should have the same size as results.
      llvm::SmallVector<std::string, 4> result_devices;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 13.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/tf2xla/transforms/legalize_tf_collective.cc

                                    DenseIntElementsAttr replica_groups,
                                    Operation* op) {
      // Use special group_key 0 to represent "all available devices". This
      // shall resolve to a DeviceAssignment that includes all devices intended for
      // replica_groups.
      IntegerAttr group_size = builder.getI32IntegerAttr(replica_groups.size());
      IntegerAttr group_key = builder.getI32IntegerAttr(0);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 16K bytes
    - Viewed (0)
  9. tensorflow/c/eager/parallel_device/parallel_device_test.cc

          TFE_ContextListDevices(context.get(), status.get()), TF_DeleteDeviceList);
      ASSERT_EQ(TF_GetCode(status.get()), TF_OK) << TF_Message(status.get());
      bool has_tpu = false;
      for (int device_index = 0; device_index < TF_DeviceListCount(devices.get());
           ++device_index) {
        std::string device_type =
            TF_DeviceListType(devices.get(), device_index, status.get());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jul 08 23:47:35 UTC 2021
    - 29.3K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/partially_decluster_pass_test.cc

      TF_ASSERT_OK(s.ToGraph(graph.get()));
    
      // This is needed to register the XLA_GPU device.
      std::vector<std::unique_ptr<Device>> devices;
      TF_ASSERT_OK(DeviceFactory::AddDevices(
          SessionOptions(), "/job:localhost/replica:0/task:0", &devices));
    
      // Scope::ToGraph loses the assigned device name since it goes through
      // GraphDef/NodeDef which does not have a field for the assigned device name.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Jun 10 12:32:39 UTC 2022
    - 23K bytes
    - Viewed (0)
Back to top