Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for genAllocator (0.37 sec)

  1. tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc

        AllocatorAttributes host_alloc_attr;
        host_alloc_attr.set_on_host(true);
        host_allocator_ = device_->GetAllocator(host_alloc_attr);
    
        AllocatorAttributes device_alloc_attr;
        device_alloc_attr.set_on_host(false);
        device_allocator_ = device_->GetAllocator(device_alloc_attr);
      }
    
     protected:
      std::unique_ptr<Device> device_;
      Allocator* host_allocator_;
      Allocator* device_allocator_;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  2. tensorflow/compiler/jit/kernels/xla_ops.cc

          for (int i = 0; i < resources.size(); i++) {
            resource_var_ptrs[resources[i]] = variable_infos[i].var()->tensor();
          }
    
          std::shared_ptr<se::DeviceMemoryAllocator> allocator =
              GetAllocator(ctx->device(), GetStream(ctx), platform_info);
          XlaComputationLaunchContext launch_context =
              GetLaunchContext(platform_info, ctx, client, allocator.get());
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 41.4K bytes
    - Viewed (0)
  3. tensorflow/compiler/jit/xla_device.cc

          .set_allowed_devices(allowed_devices_)
          .set_intra_op_parallelism_threads(intra_op_parallelism_threads_);
      return xla::ClientLibrary::GetOrCreateLocalClient(options);
    }
    
    Allocator* XlaDevice::GetAllocator(AllocatorAttributes attr) {
      mutex_lock lock(mu_);
      return GetAllocatorLocked(attr);
    }
    
    Allocator* XlaDevice::GetAllocatorLocked(AllocatorAttributes attr) {
      if (attr.on_host()) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 21:05:42 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/jit/xla_launch_util.cc

        const std::map<int, const Tensor*>& resource_vars) {
      se::Stream* stream =
          ctx->op_device_context() ? ctx->op_device_context()->stream() : nullptr;
      Allocator* allocator = ctx->device()->GetAllocator({});
    
      // Computation output should always be a tuple.
      VLOG(2) << "Result tuple shape: " << output.on_host_shape().DebugString();
      VLOG(2) << "Result tuple shape (on device): "
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 16 00:36:08 UTC 2024
    - 40.4K bytes
    - Viewed (0)
Back to top