- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 317 for Allocate (0.26 sec)
-
pkg/kubelet/cm/memorymanager/fake_memory_manager.go
klog.InfoS("Start()") return nil } func (m *fakeManager) Policy() Policy { klog.InfoS("Policy()") return NewPolicyNone() } func (m *fakeManager) Allocate(pod *v1.Pod, container *v1.Container) error { klog.InfoS("Allocate", "pod", klog.KObj(pod), "containerName", container.Name) return nil } func (m *fakeManager) AddContainer(pod *v1.Pod, container *v1.Container, containerID string) {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Sep 27 13:02:15 UTC 2023 - 3.1K bytes - Viewed (0) -
src/runtime/cgo/asm_s390x.s
TEXT crosscall2(SB),NOSPLIT|NOFRAME,$0 // Start with standard C stack frame layout and linkage. // Save R6-R15 in the register save area of the calling function. STMG R6, R15, 48(R15) // Allocate 96 bytes on the stack. MOVD $-96(R15), R15 // Save F8-F15 in our stack frame. FMOVD F8, 32(R15) FMOVD F9, 40(R15) FMOVD F10, 48(R15) FMOVD F11, 56(R15) FMOVD F12, 64(R15) FMOVD F13, 72(R15)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 12 00:43:51 UTC 2023 - 1.7K bytes - Viewed (0) -
test/fixedbugs/issue67255.go
for i := 10; i < 200; i++ { // The objects we're allocating here are pointer-ful. Some will // max out their size class, which are the ones we want. // We also allocate from small to large, so that the object which // maxes out its size class is the last one allocated in that class. // This allocation pattern leaves the next object in the class // unallocated, which we need to reproduce the bug. objs = append(objs, make([]*byte, i))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:09:06 UTC 2024 - 922 bytes - Viewed (0) -
src/runtime/os_wasm.go
//go:nosplit func osyield_no_g() { osyield() } type sigset struct{} // Called to initialize a new m (including the bootstrap m). // Called on the parent thread (main thread in case of bootstrap), can allocate memory. func mpreinit(mp *m) { mp.gsignal = malg(32 * 1024) mp.gsignal.m = mp } //go:nosplit func usleep_no_g(usec uint32) { usleep(usec) } //go:nosplit func sigsave(p *sigset) { }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 3.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfr/ir/tfr_types.h
static TFRTypeStorage* construct(TypeStorageAllocator& allocator, KeyTy key) { // Allocate a new storage instance. auto byteSize = TFRTypeStorage::totalSizeToAlloc<StringAttr>(key.size()); auto rawMem = allocator.allocate(byteSize, alignof(TFRTypeStorage)); auto result = ::new (rawMem) TFRTypeStorage(key.size());
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Dec 05 07:17:01 UTC 2023 - 4.1K bytes - Viewed (0) -
platforms/documentation/docs/src/snippets/native-binaries/cunit/groovy/libs/cunit/2.1-2/include/CUnit/MyMem.h
void* CU_realloc(void *ptr, size_t size, unsigned int uiLine, const char* szFileName); CU_EXPORT void CU_dump_memory_usage(const char*); /** c-allocate with memory tracking. */ #define CU_CALLOC(x, y) CU_calloc((x), (y), __LINE__, __FILE__) /** m-allocate with memory tracking. */ #define CU_MALLOC(x) CU_malloc((x), __LINE__, __FILE__) /** Free with memory tracking. */
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Mon Nov 27 17:53:42 UTC 2023 - 4K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/util/flowcontrol/request/list_work_estimator.go
numStored, err := e.countGetterFn(key(requestInfo)) switch { case err == ObjectCountStaleErr: // object count going stale is indicative of degradation, so we should // be conservative here and allocate maximum seats to this list request. // NOTE: if a CRD is removed, its count will go stale first and then the // pruner will eventually remove the CRD from the cache. return WorkEstimate{InitialSeats: maxSeats}
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 10 11:56:42 UTC 2024 - 7K bytes - Viewed (0) -
src/runtime/mcentral.go
traceRelease(trace) } // If we sweep spanBudget spans without finding any free // space, just allocate a fresh span. This limits the amount // of time we can spend trying to find free space and // amortizes the cost of small object sweeping over the // benefit of having a full free span to allocate from. By // setting this to 100, we limit the space overhead to 1%. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
pkg/registry/core/service/ipallocator/interfaces.go
import ( "errors" "fmt" "net" api "k8s.io/kubernetes/pkg/apis/core" ) // Interface manages the allocation of IP addresses out of a range. Interface // should be threadsafe. type Interface interface { Allocate(net.IP) error AllocateNext() (net.IP, error) Release(net.IP) error ForEach(func(net.IP)) CIDR() net.IPNet IPFamily() api.IPFamily Has(ip net.IP) bool Destroy() EnableMetrics()
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Oct 31 21:05:04 UTC 2023 - 1.6K bytes - Viewed (0) -
okhttp-testing-support/src/main/kotlin/okhttp3/FakeDns.kt
} fun assertRequests(vararg expectedHosts: String?) { assertThat(requestedHosts).containsExactly(*expectedHosts) requestedHosts.clear() } /** Allocates and returns `count` fake IPv4 addresses like [255.0.0.100, 255.0.0.101]. */ fun allocate(count: Int): List<InetAddress> { val from = nextAddress nextAddress += count return (from until nextAddress) .map {
Registered: Sun Jun 16 04:42:17 UTC 2024 - Last Modified: Mon Jan 08 01:13:22 UTC 2024 - 2.6K bytes - Viewed (0)