Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for memState (0.18 sec)

  1. src/runtime/mgc.go

    	work.tEnd = now
    	atomic.Store64(&memstats.last_gc_unix, uint64(unixNow)) // must be Unix time to make sense to user
    	atomic.Store64(&memstats.last_gc_nanotime, uint64(now)) // monotonic time for us
    	memstats.pause_ns[memstats.numgc%uint32(len(memstats.pause_ns))] = uint64(work.pauseNS)
    	memstats.pause_end[memstats.numgc%uint32(len(memstats.pause_end))] = uint64(unixNow)
    	memstats.pause_total_ns += uint64(work.pauseNS)
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
  2. src/runtime/mheap.go

    	h.cachealloc.init(unsafe.Sizeof(mcache{}), nil, nil, &memstats.mcache_sys)
    	h.specialfinalizeralloc.init(unsafe.Sizeof(specialfinalizer{}), nil, nil, &memstats.other_sys)
    	h.specialprofilealloc.init(unsafe.Sizeof(specialprofile{}), nil, nil, &memstats.other_sys)
    	h.specialReachableAlloc.init(unsafe.Sizeof(specialReachable{}), nil, nil, &memstats.other_sys)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 78K bytes
    - Viewed (0)
  3. src/runtime/malloc.go

    		if l2[ri.l2()] != nil {
    			throw("arena already initialized")
    		}
    		var r *heapArena
    		r = (*heapArena)(h.heapArenaAlloc.alloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys))
    		if r == nil {
    			r = (*heapArena)(persistentalloc(unsafe.Sizeof(*r), goarch.PtrSize, &memstats.gcMiscSys))
    			if r == nil {
    				throw("out of memory allocating heap arena metadata")
    			}
    		}
    
    		// Register the arena in allArenas if requested.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 59.6K bytes
    - Viewed (0)
  4. pkg/kubelet/eviction/helpers.go

    }
    
    // memoryUsage converts working set into a resource quantity.
    func memoryUsage(memStats *statsapi.MemoryStats) *resource.Quantity {
    	if memStats == nil || memStats.WorkingSetBytes == nil {
    		return &resource.Quantity{Format: resource.BinarySI}
    	}
    	usage := int64(*memStats.WorkingSetBytes)
    	return resource.NewQuantity(usage, resource.BinarySI)
    }
    
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Nov 01 18:46:33 UTC 2023
    - 53.6K bytes
    - Viewed (0)
  5. src/runtime/mgcscavenge.go

    				nbytes := int64(npages * pageSize)
    				gcController.heapReleased.add(nbytes)
    				gcController.heapFree.add(-nbytes)
    
    				stats := memstats.heapStats.acquire()
    				atomic.Xaddint64(&stats.committed, -nbytes)
    				atomic.Xaddint64(&stats.released, nbytes)
    				memstats.heapStats.release()
    			}
    
    			// Relock the heap, because now we need to make these pages
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:48:45 UTC 2024
    - 52.3K bytes
    - Viewed (0)
  6. src/runtime/mgcpacer.go

    	lastHeapGoal uint64
    
    	// heapLive is the number of bytes considered live by the GC.
    	// That is: retained by the most recent GC plus allocated
    	// since then. heapLive ≤ memstats.totalAlloc-memstats.totalFree, since
    	// heapAlloc includes unmarked objects that have not yet been swept (and
    	// hence goes up as we allocate and down as we sweep) while heapLive
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 55.4K bytes
    - Viewed (0)
  7. src/runtime/mprof.go

    	case memProfile:
    		size += unsafe.Sizeof(memRecord{})
    	case blockProfile, mutexProfile:
    		size += unsafe.Sizeof(blockRecord{})
    	}
    
    	b := (*bucket)(persistentalloc(size, 0, &memstats.buckhash_sys))
    	b.typ = typ
    	b.nstk = uintptr(nstk)
    	return b
    }
    
    // stk returns the slice in b holding the stack. The caller can asssume that the
    // backing array is immutable.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 53.3K bytes
    - Viewed (0)
  8. src/runtime/mbitmap.go

    // The resulting bitvector will have no more than size/goarch.PtrSize bits.
    func progToPointerMask(prog *byte, size uintptr) bitvector {
    	n := (size/goarch.PtrSize + 7) / 8
    	x := (*[1 << 30]byte)(persistentalloc(n+1, 1, &memstats.buckhash_sys))[:n+1]
    	x[len(x)-1] = 0xa1 // overflow check sentinel
    	n = runGCProg(prog, &x[0])
    	if x[len(x)-1] != 0xa1 {
    		throw("progToPointerMask: overflow")
    	}
    	return bitvector{int32(n), &x[0]}
    }
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 60K bytes
    - Viewed (0)
Back to top