Search Options

Results per page
Sort
Preferred Languages
Advance

Results 61 - 70 of 2,544 for Devices (0.13 sec)

  1. tensorflow/compiler/jit/flags.cc

                "Switch a device into 'on-demand' mode, where instead of "
                "autoclustering ops are compiled one by one just-in-time."),
    
           Flag("tf_xla_enable_xla_devices",
                &device_flags->tf_xla_enable_xla_devices,
                "Generate XLA_* devices, where placing a computation on such a "
                "device"
                "forces compilation by XLA. Deprecated."),
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 17 18:52:57 UTC 2024
    - 24.5K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/host_runtime/testdata/malformed_cluster.mlir

    module attributes {tf.versions = {producer = 888 : i32}, tf.devices = ["/job:worker/replica:0/task:0/device:CPU:0", "/job:worker/replica:0/task:0/device:TPU_SYSTEM:0", "/job:worker/replica:0/task:0/device:TPU:0"]} {
      func.func @missing_num_cores_per_replica() {
        // expected-error@+1 {{requires attribute 'num_cores_per_replica'}}
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Oct 17 16:43:16 UTC 2023
    - 742 bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/tfrt/tests/runtime_lowering_gpu.mlir

    // RUN: tf-tfrt-opt -tfrt-lower-cluster-to-runtime-ops-non-tpu -split-input-file -verify-diagnostics %s | FileCheck %s
    
    module attributes {tf.versions = {producer = 888 : i32}, tf.devices = ["/job:worker/replica:0/task:0/device:CPU:0", "/job:worker/replica:0/task:0/device:TPU_SYSTEM:0", "/job:worker/replica:0/task:0/device:GPU:0"]} {
    
      // CHECK-LABEL: @converts_cluster
      func.func @converts_cluster() {
        // CHECK: "tf.XlaLaunch"()
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Oct 13 17:41:44 UTC 2023
    - 840 bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_device_context.h

      }
    
      // Returns a device-to-device stream, in round-robin fashion.
      se::Stream* GetDeviceToDeviceStream();
    
      Status ThenExecute(Device* device, stream_executor::Stream* stream,
                         std::function<void()> func) override;
    
     private:
      bool UseMultipleStreams() const { return stream_ != host_to_device_stream_; }
    
      // The main compute stream of the device, used to synchronize the transfer
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Sep 06 19:12:29 UTC 2023
    - 5.1K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/increase_dynamism_for_auto_jit_pass_test.cc

                                      std::unique_ptr<Graph>* result) {
      std::vector<std::unique_ptr<Device>> devices;
      devices.push_back(FakeDevice::Make(kDeviceName, DEVICE_GPU));
      devices.push_back(FakeDevice::Make(kHostName, DEVICE_CPU));
    
      std::unique_ptr<DeviceSet> device_set(new DeviceSet());
      for (auto& device : devices) {
        device_set->AddDevice(device.get());
      }
    
      auto graph = std::make_unique<Graph>(OpRegistry::Global());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 18.3K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/mark_for_compilation_pass.cc

        const Cluster& cluster_a, const Cluster& cluster_b) {
      DeviceSet devices = cluster_a.devices();
      devices.UnionWith(cluster_b.devices());
    
      TF_ASSIGN_OR_RETURN(
          std::optional<jit::DeviceId> maybe_chosen_device,
          MaybePickDeviceForXla(device_info_cache_, devices,
                                /*allow_mixing_unknown_and_cpu=*/false));
      if (!maybe_chosen_device.has_value()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 12:19:41 UTC 2024
    - 85.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tf2xla/internal/passes/xla_broadcast.cc

    LogicalResult GetTpuDeviceAssignment(
        ClusterOp cluster, ReplicateOp replicate, mlir::ModuleOp module,
        absl::StatusOr<TPUDeviceAssignment>& status_or_tpu_device_assignment) {
      mlir::TF::RuntimeDevices devices;
      if (failed(tensorflow::GetDevicesFromOp(module, &devices))) return failure();
    
      uint32_t num_replicas = replicate.getN();
    
      auto num_cores_per_replica_attr = cluster->getAttrOfType<mlir::IntegerAttr>(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 13 18:52:07 UTC 2024
    - 13.9K bytes
    - Viewed (0)
  8. pkg/kubelet/cm/doc.go

    // to manage containers. For example, they contain functions to configure containers' cgroups,
    // ensure containers run with the desired QoS, and allocate compute resources like cpus, memory,
    // devices...
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Thu Oct 05 18:20:51 UTC 2023
    - 934 bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/tf_to_corert_pipeline_refvar.mlir

    // CHECK-NEXT: [[o_chain:%.*]], [[o:%.*]] = tfrt_fallback_async.executeop.seq([[in_chain]]) key(0) cost({{.*}}) device("/job:localhost/replica:0/task:0/device:CPU:0") "tf.VarHandleOp"()
    // CHECK-NEXT: [[o_chain_0:%.*]], [[o1:%.*]] = tfrt_fallback_async.executeop.seq([[in_chain]]) key(1) cost({{.*}}) device("/job:localhost/replica:0/task:0/device:CPU:0") "tf.ReadVariableOp"([[o]]) {dtype = f32} : 1
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 00:18:59 UTC 2024
    - 1.6K bytes
    - Viewed (0)
  10. pkg/volume/util/hostutil/hostutil.go

    )
    
    // HostUtils defines the set of methods for interacting with paths on a host.
    type HostUtils interface {
    	// DeviceOpened determines if the device (e.g. /dev/sdc) is in use elsewhere
    	// on the system, i.e. still mounted.
    	DeviceOpened(pathname string) (bool, error)
    	// PathIsDevice determines if a path is a device.
    	PathIsDevice(pathname string) (bool, error)
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Feb 28 13:38:40 UTC 2024
    - 4.2K bytes
    - Viewed (0)
Back to top