- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 1,897 for Devices (0.12 sec)
-
tensorflow/compiler/jit/xla_gpu_device.cc
auto device = std::make_unique<XlaDevice>(session_options, options); Status status = device->UseAcceleratorDeviceInfo(); if (!status.ok()) { LOG(INFO) << "Ignoring visible " << DEVICE_GPU_XLA_JIT << " device. Device number is " << i << ", reason: " << status; continue; } devices->push_back(std::move(device)); } return absl::OkStatus(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 6.6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util.h
// A TPU device for execution alongside its associated host CPU device. struct TPUDeviceAndHost { TPUDeviceAndHost() = default; TPUDeviceAndHost(llvm::StringRef device, llvm::StringRef host) : device(device), host(host) {} std::string device; std::string host; }; // TPU devices to be used for execution (e.g. devices for TPUExecute ops) and
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 26 09:37:10 UTC 2024 - 11.3K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/topology_hints.go
// request size and the list of available devices. deviceHints[resource] = m.generateDeviceTopologyHints(resource, available, sets.Set[string]{}, requested) } return deviceHints } func (m *ManagerImpl) deviceHasTopologyAlignment(resource string) bool { // If any device has Topology NUMANodes available, we assume they care about alignment. for _, device := range m.allDevices[resource] {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Jan 27 02:10:25 UTC 2024 - 9.9K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_cpu_device.cc
public: Status ListPhysicalDevices(std::vector<string>* devices) override; Status CreateDevices(const SessionOptions& options, const string& name_prefix, std::vector<std::unique_ptr<Device>>* devices) override; }; Status XlaCpuDeviceFactory::ListPhysicalDevices(std::vector<string>* devices) { XlaDeviceFlags* flags = GetXlaDeviceFlags();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 22 08:47:20 UTC 2024 - 5.5K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_tpu_device.cc
" device number ", i); return status; } device->SetAllowsSyncOnCompletion(false); if (tpu_xla_device_failure_closes_chips_flag) { device->SetHandleDeviceErrorCallback(&tpu::TpuNodeContext::CloseTpuHost); } devices->push_back(std::move(device)); } return absl::OkStatus(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue May 28 22:53:47 UTC 2024 - 20.9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/tpu_rewrite_device_util_test.cc
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Apr 26 09:37:10 UTC 2024 - 46.8K bytes - Viewed (0) -
pkg/kubelet/apis/podresources/server_v1_test.go
{ Name: containerName, Devices: []*podresourcesapi.ContainerDevices{}, DynamicResources: []*podresourcesapi.DynamicResource{}, }, }, }, }, }, }, { desc: "pod with devices", pods: pods, devices: devs, cpus: cpus, memory: memory,
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Apr 24 18:25:29 UTC 2024 - 45.9K bytes - Viewed (0) -
tensorflow/compiler/jit/device_util.h
}; // Caches some miscellaneous information about TF devices. Thread compatible. class DeviceInfoCache { public: bool IsGpu(DeviceId device) const { return is_gpu_[device.id()]; } bool IsCpu(DeviceId device) const { return is_cpu_[device.id()]; } absl::string_view GetNameFor(DeviceId device) const { return names_[device.id()]; } absl::StatusOr<DeviceId> GetIdFor(absl::string_view name);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 15 17:18:31 UTC 2024 - 7.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/ifrt/rewrite_cluster_to_ifrt_call.mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sat Feb 17 07:28:40 UTC 2024 - 9K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tpu_device_propagation.cc
value_to_device.insert({result, device}); } } // Checks if an operation can have TPU devices propagated through. bool IsSupportedOpToSetDevice(Operation& op) { return IsSupportedExecutorOp(op) || isa<TF::IdentityOp, TF::IdentityNOp, TF::ShapeOp>(op); } // Finds nonconflicting TPU device for an operation from its operands. If an // operand has no device or a non TPU device, or if there are conflicting
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Apr 25 16:01:03 UTC 2024 - 11K bytes - Viewed (0)