Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 57 for Deallocator (0.21 sec)

  1. tensorflow/c/tf_tensor.cc

      Allocator* allocator = nullptr;
      if (arg == nullptr) {
        allocator = cpu_allocator();
      } else {
        allocator = reinterpret_cast<Allocator*>(arg);
      }
      if (LogMemory::IsEnabled() && data != nullptr) {
        LogMemory::RecordRawDeallocation(
            "TensorFlow C Api", LogMemory::EXTERNAL_TENSOR_ALLOCATION_STEP_ID, data,
            allocator, false);
      }
      allocator->DeallocateRaw(data);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sun Apr 14 21:57:32 UTC 2024
    - 11.5K bytes
    - Viewed (0)
  2. tensorflow/c/eager/c_api.cc

        void (*deallocator)(void* data, size_t len, void* arg),
        void* deallocator_arg, TF_Status* status) {
      tensorflow::Device* device = nullptr;
      tensorflow::EagerContext* context =
          tensorflow::ContextFromInterface(tensorflow::unwrap(ctx));
      status->status = context->FindDeviceFromName(device_name, &device);
      if (!status->status.ok()) {
        deallocator(data, len, deallocator_arg);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 08:11:23 UTC 2024
    - 44K bytes
    - Viewed (0)
  3. tensorflow/c/c_api_test.cc

      TF_TString_Init(&tstr[0]);
      TF_TString_Copy(&tstr[0], test_string, sizeof(test_string) - 1);
    
      auto deallocator = [](void* data, size_t len, void* arg) {};
      unique_tensor_ptr t_in(TF_NewTensor(TF_STRING, nullptr, 0, &tstr[0],
                                          sizeof(tstr), deallocator, nullptr),
                             TF_DeleteTensor);
    
      // Create a TF_Operation with the attribute t_in
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon Apr 15 03:35:10 UTC 2024
    - 96.9K bytes
    - Viewed (0)
  4. pkg/registry/core/service/portallocator/allocator.go

    	}
    
    	return a, err
    }
    
    // NewInMemory creates an in-memory allocator.
    func NewInMemory(pr net.PortRange) (*PortAllocator, error) {
    	return New(pr, func(max int, rangeSpec string, offset int) (allocator.Interface, error) {
    		return allocator.NewAllocationMapWithOffset(max, rangeSpec, offset), nil
    	})
    }
    
    // NewFromSnapshot allocates a PortAllocator and initializes it from a snapshot.
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed May 08 07:15:02 UTC 2024
    - 7.4K bytes
    - Viewed (0)
  5. pkg/registry/core/service/ipallocator/cidrallocator.go

    func (c *MetaAllocator) AllocateService(service *api.Service, ip net.IP) error {
    	allocator, err := c.getAllocator(ip)
    	if err != nil {
    		return err
    	}
    	return allocator.AllocateService(service, ip)
    }
    
    func (c *MetaAllocator) Allocate(ip net.IP) error {
    	allocator, err := c.getAllocator(ip)
    	if err != nil {
    		return err
    	}
    	return allocator.Allocate(ip)
    }
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat May 04 18:33:12 UTC 2024
    - 13.2K bytes
    - Viewed (0)
  6. tensorflow/compiler/jit/xla_host_send_recv_device_context_test.cc

        host_alloc_attr.set_on_host(true);
        host_allocator_ = device_->GetAllocator(host_alloc_attr);
    
        AllocatorAttributes device_alloc_attr;
        device_alloc_attr.set_on_host(false);
        device_allocator_ = device_->GetAllocator(device_alloc_attr);
      }
    
     protected:
      std::unique_ptr<Device> device_;
      Allocator* host_allocator_;
      Allocator* device_allocator_;
    };
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 7.2K bytes
    - Viewed (0)
  7. tensorflow/compiler/jit/xla_platform_info.cc

        // If we are on an XlaDevice, use the underlying XLA platform's allocator
        // directly. We could use the StreamExecutor's allocator which may
        // theoretically be more correct, but XLA returns a nice OOM message in a
        // Status and StreamExecutor does not.
        //
        // Importantly we can't use ctx->device()->GetAllocator() as the allocator
        // (which xla_allocator above uses) as on an XlaDevice, this is a dummy
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 17:23:27 UTC 2024
    - 17.4K bytes
    - Viewed (0)
  8. tensorflow/compiler/jit/xla_device.cc

      }
    
      Status status;
      if (alloc_attrs.on_host()) {
        *tensor = parsed;
      } else {
        Allocator* allocator;
        {
          mutex_lock lock(mu_);
          allocator = GetAllocatorLocked(alloc_attrs);
        }
        Tensor copy(allocator, parsed.dtype(), parsed.shape());
        TF_RETURN_IF_ERROR(
            device_context->CopyCPUTensorToDeviceSync(&parsed, this, &copy));
        *tensor = copy;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Mon May 20 21:05:42 UTC 2024
    - 24.3K bytes
    - Viewed (0)
  9. pkg/registry/core/service/ipallocator/controller/repairip.go

    	defer r.svcQueue.ShutDown()
    	r.broadcaster.StartRecordingToSink(stopCh)
    	defer r.broadcaster.Shutdown()
    
    	klog.Info("Starting ipallocator-repair-controller")
    	defer klog.Info("Shutting down ipallocator-repair-controller")
    
    	if !cache.WaitForNamedCacheSync("ipallocator-repair-controller", stopCh, r.ipAddressSynced, r.servicesSynced, r.serviceCIDRSynced) {
    		return
    	}
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Sat May 04 18:33:12 UTC 2024
    - 24.7K bytes
    - Viewed (0)
  10. tensorflow/compiler/jit/kernels/xla_ops.cc

          }
    
          std::shared_ptr<se::DeviceMemoryAllocator> allocator =
              GetAllocator(ctx->device(), GetStream(ctx), platform_info);
          XlaComputationLaunchContext launch_context =
              GetLaunchContext(platform_info, ctx, client, allocator.get());
    
          const xla::HloInputOutputAliasConfig& input_output_alias =
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 22:46:36 UTC 2024
    - 41.4K bytes
    - Viewed (0)
Back to top