Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 3,466 for Devices (0.13 sec)

  1. tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/tpu_variable_runtime_reformatting.cc

        device_list_for_alias.reserve(device_list.size());
    
        for (auto device : device_list)
          device_list_for_alias.emplace_back(
              mlir::cast<StringAttr>(device).getValue());
    
        devices.insert({device_alias, device_list_for_alias});
      }
    
      OpBuilder builder(replicate);
      builder.setInsertionPoint(while_op);
      // Create per-device variables for formatting state, and add them to the while
      // loop.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 21.9K bytes
    - Viewed (0)
  2. pkg/volume/util/device_util_unsupported.go

    	return portalHostMap, nil
    }
    
    // FindDevicesForISCSILun unsupported returns nil
    func (handler *deviceHandler) FindDevicesForISCSILun(targetIqn string, lun int) ([]string, error) {
    	devices := []string{}
    	return devices, nil
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Tue Aug 24 19:47:49 UTC 2021
    - 1.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/testdata/spmd.mlir

      func.func @main(%arg0: tensor<*xf32> {tf.device = "/job:localhost/replica:0/task:0/device:CPU:0"}) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Dec 12 04:22:33 UTC 2023
    - 1.5K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/utils/tpu_cluster_util.cc

        std::function<WalkResult(Operation*, tf_device::ClusterOp,
                                 std::optional<std::string>)>
            callback) {
      mlir::TF::RuntimeDevices devices;
      if (failed(tensorflow::GetDevicesFromOp(module, &devices))) return failure();
      const CallGraph call_graph(module);
      // symbol_table caches callees in the CallGraph.
      SymbolTableCollection symbol_table;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 06 04:50:13 UTC 2023
    - 4.7K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/test_util.cc

    }
    
    Device* DeviceSetup::GetDevice(const string& device_name) {
      if (device_mgr_ == nullptr) {
        return nullptr;
      }
    
      string full_device_name = absl::StrCat(
          "/job:localhost/replica:0/task:0/device:", device_name, ":0");
      Device* device;
      TF_CHECK_OK(device_mgr_->LookupDevice(full_device_name, &device));
      return device;
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Feb 09 11:36:41 UTC 2024
    - 3.7K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/ir/tf_ops_device_helper.h

    class RuntimeDevices;
    
    // Returns true if at least one GPU device is available at runtime.
    bool CanUseGpuDevice(const RuntimeDevices &devices);
    
    // Returns true if all of the GPUs available at runtime support TensorCores
    // (NVIDIA compute capability >= 7.0).
    bool CanUseTensorCores(const RuntimeDevices &devices);
    
    // Returns true if operation does not have explicit device placement that would
    // prevent it from running on GPU device.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Nov 12 21:57:12 UTC 2021
    - 1.4K bytes
    - Viewed (0)
  7. pkg/util/procfs/procfs_linux.go

    		if len(entries) == 3 && entries[1] == "devices" {
    			return strings.TrimSpace(entries[2]), nil
    		}
    	}
    	return "", fmt.Errorf("could not find devices cgroup location")
    }
    
    // GetFullContainerName gets the container name given the root process id of the container.
    // E.g. if the devices cgroup for the container is stored in /sys/fs/cgroup/devices/docker/nginx,
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Mon Jan 16 09:22:35 UTC 2023
    - 4.1K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/xla_device.h

        // The name of the compilation device (e.g., "XLA_CPU_JIT");
        string compilation_device_name;
    
        // If 'use_multiple_streams' is true, we create separate streams for
        // compute, host-to-device, and device-to-host communication.
        bool use_multiple_streams = false;
    
        // If true, the XLA devices with the same device ordinal will share the same
        // compute stream. Otherwise each XLA device will having their own compute
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 13.4K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/tests/tpu-variable-runtime-reformatting.mlir

                 %arg2: !tf_res_md_f32 {tf.device = "/device:TPU:0"},
                 %arg3: !tf_res_md_f32 {tf.device = "/device:TPU:1"}) {
    
        %0 = "tf.Const"() {value = dense<100> : tensor<i32>} : () -> tensor<i32>
        // CHECK: %[[STATE0:.*]] = "tf.VarHandleOp"()
        // CHECK-SAME: device = "/device:TPU:0"
        // CHECK: %[[STATE1:.*]] = "tf.VarHandleOp"()
        // CHECK-SAME: device = "/device:TPU:1"
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 31 08:59:10 UTC 2023
    - 25.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/tests/tf_device_ops.mlir

      tf_device.replicate() {n = 2 : i32, devices = {TPU_REPLICATED_CORE_0 = ["/DEVICE:0", "/DEVICE:1"], TPU_REPLICATED_CORE_1 = ["/DEVICE:2", "/DEVICE:3"]}} {
        tf_device.return
      }
      func.return
    
    // CHECK:      tf_device.replicate
    // CHECK-SAME: devices = {TPU_REPLICATED_CORE_0 = ["/DEVICE:0", "/DEVICE:1"], TPU_REPLICATED_CORE_1 = ["/DEVICE:2", "/DEVICE:3"]}
    // CHECK-SAME: n = 2
    // CHECK-NEXT:   tf_device.return
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jan 23 23:53:20 UTC 2024
    - 7.7K bytes
    - Viewed (0)
Back to top