Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 1,897 for Devices (0.12 sec)

  1. tensorflow/compiler/jit/xla_gpu_device.cc

        auto device = std::make_unique<XlaDevice>(session_options, options);
    
        Status status = device->UseAcceleratorDeviceInfo();
        if (!status.ok()) {
          LOG(INFO) << "Ignoring visible " << DEVICE_GPU_XLA_JIT
                    << " device. Device number is " << i << ", reason: " << status;
          continue;
        }
    
        devices->push_back(std::move(device));
      }
      return absl::OkStatus();
    }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 6.6K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h

    // A TPU device for execution alongside its associated host CPU device.
    struct TPUDeviceAndHost {
      TPUDeviceAndHost() = default;
      TPUDeviceAndHost(llvm::StringRef device, llvm::StringRef host)
          : device(device), host(host) {}
    
      std::string device;
      std::string host;
    };
    
    // TPU devices to be used for execution (e.g. devices for TPUExecute ops) and
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 11.3K bytes
    - Viewed (0)
  3. pkg/kubelet/cm/devicemanager/topology_hints.go

    		// request size and the list of available devices.
    		deviceHints[resource] = m.generateDeviceTopologyHints(resource, available, sets.Set[string]{}, requested)
    	}
    
    	return deviceHints
    }
    
    func (m *ManagerImpl) deviceHasTopologyAlignment(resource string) bool {
    	// If any device has Topology NUMANodes available, we assume they care about alignment.
    	for _, device := range m.allDevices[resource] {
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat Jan 27 02:10:25 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_cpu_device.cc

     public:
      Status ListPhysicalDevices(std::vector<string>* devices) override;
      Status CreateDevices(const SessionOptions& options, const string& name_prefix,
                           std::vector<std::unique_ptr<Device>>* devices) override;
    };
    
    Status XlaCpuDeviceFactory::ListPhysicalDevices(std::vector<string>* devices) {
      XlaDeviceFlags* flags = GetXlaDeviceFlags();
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 22 08:47:20 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/xla_tpu_device.cc

                                  " device number ", i);
          return status;
        }
        device->SetAllowsSyncOnCompletion(false);
        if (tpu_xla_device_failure_closes_chips_flag) {
          device->SetHandleDeviceErrorCallback(&tpu::TpuNodeContext::CloseTpuHost);
        }
    
        devices->push_back(std::move(device));
      }
    
      return absl::OkStatus();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue May 28 22:53:47 UTC 2024
    - 20.9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util_test.cc

        devices.push_back(
            llvm::formatv("/job:worker/replica:0/task:{0}/device:CPU:0", task)
                .str());
        devices.push_back(
            llvm::formatv("/job:worker/replica:0/task:{0}/device:TPU_SYSTEM:0",
                          task)
                .str());
        for (int device = 0; device < num_devices_per_task; ++device)
          devices.push_back(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 46.8K bytes
    - Viewed (0)
  7. pkg/kubelet/apis/podresources/server_v1_test.go

    							{
    								Name:             containerName,
    								Devices:          []*podresourcesapi.ContainerDevices{},
    								DynamicResources: []*podresourcesapi.DynamicResource{},
    							},
    						},
    					},
    				},
    			},
    		},
    		{
    			desc:             "pod with devices",
    			pods:             pods,
    			devices:          devs,
    			cpus:             cpus,
    			memory:           memory,
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Apr 24 18:25:29 UTC 2024
    - 45.9K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/device_util.h

    };
    
    // Caches some miscellaneous information about TF devices.  Thread compatible.
    class DeviceInfoCache {
     public:
      bool IsGpu(DeviceId device) const { return is_gpu_[device.id()]; }
      bool IsCpu(DeviceId device) const { return is_cpu_[device.id()]; }
    
      absl::string_view GetNameFor(DeviceId device) const {
        return names_[device.id()];
      }
    
      absl::StatusOr<DeviceId> GetIdFor(absl::string_view name);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 15 17:18:31 UTC 2024
    - 7.1K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tfrt/tests/ifrt/rewrite_cluster_to_ifrt_call.mlir

    func.func @serving_default(%arg0: tensor<1x3xf32>) -> () {
      %outputs  =  "tf.TPUCompilationResult"() {_tpu_compilation_status = "cluster", device = ""} : () -> tensor<!tf_type.string>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Feb 17 07:28:40 UTC 2024
    - 9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/transforms/tpu_device_propagation.cc

        value_to_device.insert({result, device});
      }
    }
    
    // Checks if an operation can have TPU devices propagated through.
    bool IsSupportedOpToSetDevice(Operation& op) {
      return IsSupportedExecutorOp(op) ||
             isa<TF::IdentityOp, TF::IdentityNOp, TF::ShapeOp>(op);
    }
    
    // Finds nonconflicting TPU device for an operation from its operands. If an
    // operand has no device or a non TPU device, or if there are conflicting
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 11K bytes
    - Viewed (0)
Back to top