- Sort Score
- Result 10 results
- Languages All
Results 131 - 140 of 1,371 for Allocate (0.18 sec)
-
pkg/kubelet/cm/cpumanager/cpu_assignment.go
// NUMA nodes to allocate any 'remainder' CPUs from (in cases where the total // number of CPUs to allocate cannot be evenly distributed across the chosen // set of NUMA nodes). This "balance score" is calculated as the standard // deviation of how many CPUs will be available on each NUMA node after all // evenly distributed and remainder CPUs are allocated. The subset with the
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu Jan 25 23:56:21 UTC 2024 - 36.3K bytes - Viewed (0) -
staging/src/k8s.io/apimachinery/pkg/runtime/serializer/protobuf/protobuf.go
} prefixSize := uint64(len(s.prefix)) var unk runtime.Unknown switch t := obj.(type) { case *runtime.Unknown: estimatedSize := prefixSize + uint64(t.Size()) data := memAlloc.Allocate(estimatedSize) i, err := t.MarshalTo(data[prefixSize:]) if err != nil { return err } copy(data, s.prefix) _, err = w.Write(data[:prefixSize+uint64(i)]) return err default:
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Feb 23 13:38:23 UTC 2022 - 17.8K bytes - Viewed (0) -
src/runtime/mgcwork.go
if b != nil { b.checkempty() } } // Record that this may acquire the wbufSpans or heap lock to // allocate a workbuf. lockWithRankMayAcquire(&work.wbufSpans.lock, lockRankWbufSpans) lockWithRankMayAcquire(&mheap_.lock, lockRankMheap) if b == nil { // Allocate more workbufs. var s *mspan if work.wbufSpans.free.first != nil { lock(&work.wbufSpans.lock) s = work.wbufSpans.free.first
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 12.9K bytes - Viewed (0) -
platforms/core-runtime/base-services/src/main/java/org/gradle/internal/work/WorkerLeaseRegistry.java
package org.gradle.internal.work; import org.gradle.internal.resources.ResourceLock; /** * Used to obtain and release worker leases to run work. There are a limited number of leases available and this service is used to allocate these to worker threads. * * Used where the operation cannot be packaged as a unit of work, for example when the operation is started and completed in response to separate * events. */
Registered: Wed Jun 12 18:38:38 UTC 2024 - Last Modified: Fri Sep 22 08:48:02 UTC 2023 - 2K bytes - Viewed (0) -
src/runtime/malloc.go
// persistentChunkSize is the number of bytes we allocate when we grow // a persistentAlloc. const persistentChunkSize = 256 << 10 // persistentChunks is a list of all the persistent chunks we have // allocated. The list is maintained through the first word in the // persistent chunk. This is updated atomically. var persistentChunks *notInHeap // Wrapper around sysAlloc that can allocate small chunks.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
pilot/pkg/serviceregistry/serviceentry/store_test.go
} if !store.allocateNeeded { t.Errorf("expected allocate needed") } store.allocateNeeded = false store.deleteServices(httpDNSRR.NamespacedName()) got = store.getAllServices() if len(got) != 0 { t.Errorf("got unexpected services %v", got) } if store.allocateNeeded { t.Errorf("expected no allocate needed") } }
Registered: Fri Jun 14 15:00:06 UTC 2024 - Last Modified: Wed May 29 15:31:09 UTC 2024 - 8.1K bytes - Viewed (0) -
src/runtime/map_fast64.go
continue } insertb = b inserti = i goto done } ovf := b.overflow(t) if ovf == nil { break } b = ovf } // Did not find mapping for key. Allocate new cell & add entry. // If we hit the max load factor or we have too many overflow buckets, // and we're not already in the middle of growing, start growing.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 14.1K bytes - Viewed (0) -
src/runtime/cgo/gcc_linux_amd64.c
clang before 3.8. The test for this is misc/cgo/testsanitizers. GCC works hard to eliminate a seemingly unnecessary call to malloc, so we actually use the memory we allocate. */ setg_gcc = setg; pbounds = (uintptr*)malloc(2 * sizeof(uintptr)); if (pbounds == NULL) { fatalf("malloc failed: %s", strerror(errno)); } _cgo_set_stacklo(g, pbounds); free(pbounds);
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 22:06:46 UTC 2023 - 2.3K bytes - Viewed (0) -
tensorflow/c/tf_buffer.cc
} const size_t proto_size = in.ByteSizeLong(); void* buf = port::Malloc(proto_size); if (buf == nullptr) { return tensorflow::errors::ResourceExhausted( "Failed to allocate memory to serialize message of type '", in.GetTypeName(), "' and size ", proto_size); } if (!in.SerializeWithCachedSizesToArray(static_cast<uint8_t*>(buf))) { port::Free(buf);
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Sun Apr 14 21:57:32 UTC 2024 - 3K bytes - Viewed (0) -
src/runtime/cgo/gcc_linux_arm64.c
clang before 3.8. The test for this is misc/cgo/testsanitizers. GCC works hard to eliminate a seemingly unnecessary call to malloc, so we actually use the memory we allocate. */ setg_gcc = setg; pbounds = (uintptr*)malloc(2 * sizeof(uintptr)); if (pbounds == NULL) { fatalf("malloc failed: %s", strerror(errno)); } _cgo_set_stacklo(g, pbounds); free(pbounds);
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 22:06:46 UTC 2023 - 2.2K bytes - Viewed (0)