- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 13 for genAllocator (0.2 sec)
-
pkg/registry/core/service/ipallocator/cidrallocator.go
} } func (c *MetaAllocator) AllocateService(service *api.Service, ip net.IP) error { allocator, err := c.getAllocator(ip) if err != nil { return err } return allocator.AllocateService(service, ip) } func (c *MetaAllocator) Allocate(ip net.IP) error { allocator, err := c.getAllocator(ip) if err != nil { return err } return allocator.Allocate(ip) }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat May 04 18:33:12 UTC 2024 - 13.2K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.cc
custom_allocator); } std::shared_ptr<se::DeviceMemoryAllocator> GetAllocator( DeviceBase* device, se::Stream* stream, const XlaPlatformInfo& platform_info) { if (platform_info.custom_allocator()) { return platform_info.custom_allocator(); } auto* alloc = device->GetAllocator({}); if (!stream) { // Stream is not set for the host platform. se::Platform* platform =
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 17:23:27 UTC 2024 - 17.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc
AllocatorAttributes host_alloc_attr; host_alloc_attr.set_on_host(true); host_allocator_ = device_->GetAllocator(host_alloc_attr); AllocatorAttributes device_alloc_attr; device_alloc_attr.set_on_host(false); device_allocator_ = device_->GetAllocator(device_alloc_attr); } protected: std::unique_ptr<Device> device_; Allocator* host_allocator_; Allocator* device_allocator_;
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_compiler_options_util.cc
options.graph_def_version = function_library.graph_def_version(); options.allow_cpu_custom_calls = (platform_info.platform_id() == se::host::kHostPlatformId); options.device_allocator = GetAllocator(device, stream, platform_info); if (platform_info.xla_device_metadata()) { options.shape_determination_fns = platform_info.xla_device_metadata()->default_shape_determination_fns(); }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 6.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_platform_info.h
// // This is necessary because for XLA devices the underlying TF allocator returns // dummy tensors. // // `stream` parameter is nullable when running on host. std::shared_ptr<se::DeviceMemoryAllocator> GetAllocator( DeviceBase* device, se::Stream* stream, const XlaPlatformInfo& platform_info); } // namespace tensorflow
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 7.2K bytes - Viewed (0) -
tensorflow/compiler/jit/kernels/xla_ops.cc
for (int i = 0; i < resources.size(); i++) { resource_var_ptrs[resources[i]] = variable_infos[i].var()->tensor(); } std::shared_ptr<se::DeviceMemoryAllocator> allocator = GetAllocator(ctx->device(), GetStream(ctx), platform_info); XlaComputationLaunchContext launch_context = GetLaunchContext(platform_info, ctx, client, allocator.get());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri May 17 22:46:36 UTC 2024 - 41.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.cc
.set_allowed_devices(allowed_devices_) .set_intra_op_parallelism_threads(intra_op_parallelism_threads_); return xla::ClientLibrary::GetOrCreateLocalClient(options); } Allocator* XlaDevice::GetAllocator(AllocatorAttributes attr) { mutex_lock lock(mu_); return GetAllocatorLocked(attr); } Allocator* XlaDevice::GetAllocatorLocked(AllocatorAttributes attr) { if (attr.on_host()) {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Mon May 20 21:05:42 UTC 2024 - 24.3K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_launch_util_test.cc
host_alloc_attr.set_on_host(true); host_allocator_ = device_->GetAllocator(host_alloc_attr); // Get the device allocator. This should give us an AsyncValueAllocator. AllocatorAttributes device_alloc_attr; device_alloc_attr.set_on_host(false); device_allocator_ = device_->GetAllocator(device_alloc_attr); // Create the DeviceCompiler to help with compiling executables.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 28.8K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_compile_on_demand_op.cc
se::Stream* stream = ctx->op_device_context() ? ctx->op_device_context()->stream() : nullptr; std::shared_ptr<se::DeviceMemoryAllocator> allocator_ptr = GetAllocator(ctx->device(), stream, platform_info_); se::DeviceMemoryAllocator* allocator = allocator_ptr.get(); XlaComputationLaunchContext launch_context( client, allocator, client->default_device_ordinal(),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Feb 29 08:39:39 UTC 2024 - 13.4K bytes - Viewed (0) -
tensorflow/compiler/jit/xla_device.h
std::optional<std::set<int>> allowed_devices; }; // Creates a new XLA Device. XlaDevice(const SessionOptions& session_options, const Options& options); ~XlaDevice() override; Allocator* GetAllocator(AllocatorAttributes attr) override TF_LOCKS_EXCLUDED(mu_); void Compute(OpKernel* op_kernel, OpKernelContext* context) override; void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 09:53:30 UTC 2024 - 13.4K bytes - Viewed (0)