Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 13 for genAllocator (0.2 sec)

  1. pkg/registry/core/service/ipallocator/cidrallocator.go

    	}
    }
    
    func (c *MetaAllocator) AllocateService(service *api.Service, ip net.IP) error {
    	allocator, err := c.getAllocator(ip)
    	if err != nil {
    		return err
    	}
    	return allocator.AllocateService(service, ip)
    }
    
    func (c *MetaAllocator) Allocate(ip net.IP) error {
    	allocator, err := c.getAllocator(ip)
    	if err != nil {
    		return err
    	}
    	return allocator.Allocate(ip)
    }
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat May 04 18:33:12 UTC 2024
    - 13.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/xla_platform_info.cc

                             custom_allocator);
    }
    
    std::shared_ptr<se::DeviceMemoryAllocator> GetAllocator(
        DeviceBase* device, se::Stream* stream,
        const XlaPlatformInfo& platform_info) {
      if (platform_info.custom_allocator()) {
        return platform_info.custom_allocator();
      }
      auto* alloc = device->GetAllocator({});
      if (!stream) {
        // Stream is not set for the host platform.
        se::Platform* platform =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 17:23:27 UTC 2024
    - 17.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc

        AllocatorAttributes host_alloc_attr;
        host_alloc_attr.set_on_host(true);
        host_allocator_ = device_->GetAllocator(host_alloc_attr);
    
        AllocatorAttributes device_alloc_attr;
        device_alloc_attr.set_on_host(false);
        device_allocator_ = device_->GetAllocator(device_alloc_attr);
      }
    
     protected:
      std::unique_ptr<Device> device_;
      Allocator* host_allocator_;
      Allocator* device_allocator_;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_compiler_options_util.cc

      options.graph_def_version = function_library.graph_def_version();
      options.allow_cpu_custom_calls =
          (platform_info.platform_id() == se::host::kHostPlatformId);
      options.device_allocator = GetAllocator(device, stream, platform_info);
      if (platform_info.xla_device_metadata()) {
        options.shape_determination_fns =
            platform_info.xla_device_metadata()->default_shape_determination_fns();
      }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 6.4K bytes
    - Viewed (0)
  5. tensorflow/compiler/jit/xla_platform_info.h

    //
    // This is necessary because for XLA devices the underlying TF allocator returns
    // dummy tensors.
    //
    // `stream` parameter is nullable when running on host.
    std::shared_ptr<se::DeviceMemoryAllocator> GetAllocator(
        DeviceBase* device, se::Stream* stream,
        const XlaPlatformInfo& platform_info);
    
    }  // namespace tensorflow
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/kernels/xla_ops.cc

          for (int i = 0; i < resources.size(); i++) {
            resource_var_ptrs[resources[i]] = variable_infos[i].var()->tensor();
          }
    
          std::shared_ptr<se::DeviceMemoryAllocator> allocator =
              GetAllocator(ctx->device(), GetStream(ctx), platform_info);
          XlaComputationLaunchContext launch_context =
              GetLaunchContext(platform_info, ctx, client, allocator.get());
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 41.4K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/xla_device.cc

          .set_allowed_devices(allowed_devices_)
          .set_intra_op_parallelism_threads(intra_op_parallelism_threads_);
      return xla::ClientLibrary::GetOrCreateLocalClient(options);
    }
    
    Allocator* XlaDevice::GetAllocator(AllocatorAttributes attr) {
      mutex_lock lock(mu_);
      return GetAllocatorLocked(attr);
    }
    
    Allocator* XlaDevice::GetAllocatorLocked(AllocatorAttributes attr) {
      if (attr.on_host()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 21:05:42 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/xla_launch_util_test.cc

        host_alloc_attr.set_on_host(true);
        host_allocator_ = device_->GetAllocator(host_alloc_attr);
    
        // Get the device allocator. This should give us an AsyncValueAllocator.
        AllocatorAttributes device_alloc_attr;
        device_alloc_attr.set_on_host(false);
        device_allocator_ = device_->GetAllocator(device_alloc_attr);
    
        // Create the DeviceCompiler to help with compiling executables.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 28.8K bytes
    - Viewed (0)
  9. tensorflow/compiler/jit/xla_compile_on_demand_op.cc

      se::Stream* stream =
          ctx->op_device_context() ? ctx->op_device_context()->stream() : nullptr;
      std::shared_ptr<se::DeviceMemoryAllocator> allocator_ptr =
          GetAllocator(ctx->device(), stream, platform_info_);
      se::DeviceMemoryAllocator* allocator = allocator_ptr.get();
      XlaComputationLaunchContext launch_context(
          client, allocator, client->default_device_ordinal(),
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Feb 29 08:39:39 UTC 2024
    - 13.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/xla_device.h

        std::optional<std::set<int>> allowed_devices;
      };
    
      // Creates a new XLA Device.
      XlaDevice(const SessionOptions& session_options, const Options& options);
      ~XlaDevice() override;
    
      Allocator* GetAllocator(AllocatorAttributes attr) override
          TF_LOCKS_EXCLUDED(mu_);
      void Compute(OpKernel* op_kernel, OpKernelContext* context) override;
      void ComputeAsync(AsyncOpKernel* op_kernel, OpKernelContext* context,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 21 09:53:30 UTC 2024
    - 13.4K bytes
    - Viewed (0)
Back to top