- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 44 for fallocate (0.14 sec)
-
src/runtime/mheap.go
// to safely deal with potentially invalid pointers, since resolving // such pointers may race with a span being allocated. type mSpanState uint8 const ( mSpanDead mSpanState = iota mSpanInUse // allocated for garbage collected heap mSpanManual // allocated for manual management (e.g., stack allocator) ) // mSpanStateNames are the names of the span states, indexed by // mSpanState.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/runtime/malloc.go
// persistentChunkSize is the number of bytes we allocate when we grow // a persistentAlloc. const persistentChunkSize = 256 << 10 // persistentChunks is a list of all the persistent chunks we have // allocated. The list is maintained through the first word in the // persistent chunk. This is updated atomically. var persistentChunks *notInHeap // Wrapper around sysAlloc that can allocate small chunks.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
pkg/scheduler/framework/plugins/dynamicresources/dynamicresources.go
clearAllocation := state.informationsForClaim[index].structuredParameters // Before we tell a driver to deallocate a claim, we // have to stop telling it to allocate. Otherwise, // depending on timing, it will deallocate the claim, // see a PodSchedulingContext with selected node, and // allocate again for that same node. if !clearAllocation && state.podSchedulingState.schedulingCtx != nil &&
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 30 15:22:37 UTC 2024 - 75.9K bytes - Viewed (0) -
pkg/kubelet/cm/devicemanager/manager_test.go
as.Equal(1, len(testManager.endpoints)) // Stops resourceName2 endpoint. Verifies its stopTime is set, allocate and // preStartContainer calls return errors. e2.client.Disconnect() as.False(e2.stopTime.IsZero()) _, err = e2.allocate([]string{"Device1"}) reflect.DeepEqual(err, fmt.Errorf(errEndpointStopped, e2)) _, err = e2.preStartContainer([]string{"Device1"})
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 06:25:43 UTC 2024 - 65K bytes - Viewed (0) -
pkg/scheduler/framework/plugins/dynamicresources/dynamicresources_test.go
want: want{ prefilter: result{ status: framework.NewStatus(framework.Skip), }, postfilter: result{ status: framework.NewStatus(framework.Unschedulable, `no new claims to deallocate`), }, }, }, "claim-reference": { pod: podWithClaimName, claims: []*resourcev1alpha2.ResourceClaim{allocatedClaim, otherClaim}, want: want{ prebind: result{
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Mon May 27 07:57:10 UTC 2024 - 61.9K bytes - Viewed (0) -
src/runtime/mgcmark.go
if d.fn != nil { // Scan the func value, which could be a stack allocated closure. // See issue 30453. scanblock(uintptr(unsafe.Pointer(&d.fn)), goarch.PtrSize, &oneptrmask[0], gcw, &state) } if d.link != nil { // The link field of a stack-allocated defer record might point // to a heap-allocated defer record. Keep that heap record live.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 21:25:11 UTC 2024 - 52.5K bytes - Viewed (0) -
src/runtime/mprof.go
// elsewhere. var disableMemoryProfiling bool // A MemProfileRecord describes the live objects allocated // by a particular call sequence (stack trace). type MemProfileRecord struct { AllocBytes, FreeBytes int64 // number of bytes allocated, freed AllocObjects, FreeObjects int64 // number of objects allocated, freed Stack0 [32]uintptr // stack trace for this record; ends at first 0 entry }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0) -
src/runtime/map.go
// For hint < 0 overLoadFactor returns false since hint < bucketCnt. B := uint8(0) for overLoadFactor(hint, B) { B++ } h.B = B // allocate initial hash table // if B == 0, the buckets field is allocated lazily later (in mapassign) // If hint is large zeroing this memory could take a while. if h.B != 0 { var nextOverflow *bmap h.buckets, nextOverflow = makeBucketArray(t, h.B, nil)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 57.6K bytes - Viewed (0) -
src/runtime/mgcscavenge.go
// The background scavenger and heap-growth scavenger only release memory in chunks // that have not been densely-allocated for at least 1 full GC cycle. The reason // behind this is likelihood of reuse: the Go heap is allocated in a first-fit order // and by the end of the GC mark phase, the heap tends to be densely packed. Releasing
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:48:45 UTC 2024 - 52.3K bytes - Viewed (0) -
src/runtime/mgc.go
// goroutine needs another span, it first attempts to reclaim that much memory // by sweeping. When a goroutine needs to allocate a new small-object span, it // sweeps small-object spans for the same object size until it frees at least // one object. When a goroutine needs to allocate large-object span from heap, // it sweeps spans until it frees at least that many pages into heap. There is
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0)