- Sort Score
- Result 10 results
- Languages All
Results 41 - 50 of 75 for memState (1.09 sec)
-
src/runtime/testdata/testprog/lockosthread.go
// <4 core machines, we are still at the whim of the kernel scheduler. runtime.GOMAXPROCS(4) go func() { // Stop the world; race with LockOSThread below. var m runtime.MemStats for { runtime.ReadMemStats(&m) } }() // Try to synchronize both LockOSThreads. start := time.Now().Add(10 * time.Millisecond) var wg sync.WaitGroup wg.Add(2)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Mar 21 20:00:09 UTC 2024 - 6.6K bytes - Viewed (0) -
src/runtime/arena.go
// won't change out from under us). stats := memstats.heapStats.acquire() atomic.Xaddint64(&stats.committed, -int64(s.npages*pageSize)) atomic.Xaddint64(&stats.inHeap, -int64(s.npages*pageSize)) atomic.Xadd64(&stats.largeFreeCount, 1) atomic.Xadd64(&stats.largeFree, int64(s.elemsize)) memstats.heapStats.release() // This counts as a free, so update heapLive.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:44:56 UTC 2024 - 37.9K bytes - Viewed (0) -
src/runtime/malloc.go
if l2[ri.l2()] != nil { throw("arena already initialized") } var r *heapArena r = (*heapArena)(h.heapArenaAlloc.alloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys)) if r == nil { r = (*heapArena)(persistentalloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys)) if r == nil { throw("out of memory allocating heap arena metadata") } } // Register the arena in allArenas if requested.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/metrics_test.go
case "/gc/heap/tiny/allocs:objects": // Currently, MemStats adds tiny alloc count to both Mallocs AND Frees. // The reason for this is because MemStats couldn't be extended at the time // but there was a desire to have Mallocs at least be a little more representative, // while having Mallocs - Frees still represent a live object count. // Unfortunately, MemStats doesn't actually export a large allocation count,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 45K bytes - Viewed (0) -
src/runtime/mgcsweep.go
// wasn't totally filled, but then swept, still has all of its // free slots zeroed. s.needzero = 1 stats := memstats.heapStats.acquire() atomic.Xadd64(&stats.smallFreeCount[spc.sizeclass()], int64(nfreed)) memstats.heapStats.release() // Count the frees in the inconsistent, internal stats. gcController.totalFree.Add(int64(nfreed) * int64(s.elemsize)) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:52:18 UTC 2024 - 32.9K bytes - Viewed (0) -
pkg/kubelet/eviction/helpers.go
} // memoryUsage converts working set into a resource quantity. func memoryUsage(memStats *statsapi.MemoryStats) *resource.Quantity { if memStats == nil || memStats.WorkingSetBytes == nil { return &resource.Quantity{Format: resource.BinarySI} } usage := int64(*memStats.WorkingSetBytes) return resource.NewQuantity(usage, resource.BinarySI) }
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Nov 01 18:46:33 UTC 2023 - 53.6K bytes - Viewed (0) -
cmd/admin-handlers.go
}(), PoolNumbers: server.PoolNumbers, MemStats: madmin.MemStats{ Alloc: server.MemStats.Alloc, TotalAlloc: server.MemStats.TotalAlloc, Mallocs: server.MemStats.Mallocs, Frees: server.MemStats.Frees, HeapAlloc: server.MemStats.HeapAlloc, }, GoMaxProcs: server.GoMaxProcs, NumCPU: server.NumCPU,
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 98K bytes - Viewed (0) -
src/runtime/stack_test.go
const ( BatchSize = 32 BatchCount = 256 ArraySize = 1024 RecursionDepth = 128 ) if testing.Short() { return } defer GOMAXPROCS(GOMAXPROCS(BatchSize)) s0 := new(MemStats) ReadMemStats(s0) for b := 0; b < BatchCount; b++ { c := make(chan bool, BatchSize) for i := 0; i < BatchSize; i++ { go func() { var f func(k int, a [ArraySize]byte)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 14 00:03:57 UTC 2023 - 23.1K bytes - Viewed (0) -
src/runtime/os_darwin.go
// Find out that stack size for our own stack guard. if pthread_attr_getstacksize(&attr, &stacksize) != 0 { writeErrStr(failthreadcreate) exit(1) } g0.stack.hi = stacksize // for mstart memstats.stacks_sys.add(int64(stacksize)) // Tell the pthread library we won't join with this thread. if pthread_attr_setdetachstate(&attr, _PTHREAD_CREATE_DETACHED) != 0 { writeErrStr(failthreadcreate) exit(1) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Dec 05 20:34:30 UTC 2023 - 11.9K bytes - Viewed (0) -
src/runtime/mgclimit.go
if change > 0 && headroom <= uint64(change) { l.overflow += uint64(change) - headroom l.bucket.fill = l.bucket.capacity if !enabled { l.enabled.Store(true) l.lastEnabledCycle.Store(memstats.numgc + 1) } return } // Handle non-limiting cases. if change < 0 && l.bucket.fill <= uint64(-change) { // Bucket emptied. l.bucket.fill = 0 } else { // All other cases.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 22 22:07:41 UTC 2024 - 17.3K bytes - Viewed (0)