- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 38 for memState (0.18 sec)
-
src/runtime/export_test.go
metricsUnlock() startTheWorld(stw) } var DoubleCheckReadMemStats = &doubleCheckReadMemStats // ReadMemStatsSlow returns both the runtime-computed MemStats and // MemStats accumulated by scanning the heap. func ReadMemStatsSlow() (base, slow MemStats) { stw := stopTheWorld(stwForTestReadMemStatsSlow) // Run on the system stack to avoid stack growth allocation. systemstack(func() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:50:53 UTC 2024 - 46.1K bytes - Viewed (0) -
cmd/handler-api.go
available = (limit * 9) / 10 return } } // for all other platforms limits are based on virtual memory. memStats, err := mem.VirtualMemory() if err != nil { return } // A valid value is available return its 90% available = (memStats.Available * 9) / 10 return } func (t *apiConfig) init(cfg api.Config, setDriveCounts []int, legacy bool) { t.mu.Lock()
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Wed Jun 12 08:13:12 UTC 2024 - 10.4K bytes - Viewed (0) -
src/cmd/compile/internal/syntax/parser_test.go
} var skipRx *regexp.Regexp if *skip != "" { var err error skipRx, err = regexp.Compile(*skip) if err != nil { t.Fatalf("invalid argument for -skip (%v)", err) } } var m1 runtime.MemStats runtime.ReadMemStats(&m1) start := time.Now() type parseResult struct { filename string lines uint } goroot := testenv.GOROOT(t) results := make(chan parseResult) go func() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Mar 14 16:30:19 UTC 2024 - 12.2K bytes - Viewed (0) -
src/runtime/mspanset.go
if spineLen == b.spineCap { // Grow the spine. newCap := b.spineCap * 2 if newCap == 0 { newCap = spanSetInitSpineCap } newSpine := persistentalloc(newCap*goarch.PtrSize, cpu.CacheLineSize, &memstats.gcMiscSys) if b.spineCap != 0 { // Blocks are allocated off-heap, so // no write barriers. memmove(newSpine, spine.p, b.spineCap*goarch.PtrSize) } spine = spanSetSpinePointer{newSpine}
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 13.1K bytes - Viewed (0) -
src/image/gif/reader_test.go
hugeGIF.Image[i] = img hugeGIF.Delay[i] = 60 } buf := new(bytes.Buffer) if err := EncodeAll(buf, hugeGIF); err != nil { t.Fatal("EncodeAll:", err) } s0, s1 := new(runtime.MemStats), new(runtime.MemStats) runtime.GC() defer debug.SetGCPercent(debug.SetGCPercent(5)) runtime.ReadMemStats(s0) if _, err := Decode(buf); err != nil { t.Fatal("Decode:", err) } runtime.ReadMemStats(s1)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 11 16:15:54 UTC 2024 - 13.3K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/compile.go
continue } f.pass = &p phaseName = p.name if f.Log() { f.Logf(" pass %s begin\n", p.name) } // TODO: capture logging during this pass, add it to the HTML var mStart runtime.MemStats if logMemStats || p.mem { runtime.ReadMemStats(&mStart) } if checkEnabled && !f.scheduled { // Test that we don't depend on the value order, by randomizing
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 22 14:55:18 UTC 2024 - 18.6K bytes - Viewed (0) -
src/runtime/arena.go
// won't change out from under us). stats := memstats.heapStats.acquire() atomic.Xaddint64(&stats.committed, -int64(s.npages*pageSize)) atomic.Xaddint64(&stats.inHeap, -int64(s.npages*pageSize)) atomic.Xadd64(&stats.largeFreeCount, 1) atomic.Xadd64(&stats.largeFree, int64(s.elemsize)) memstats.heapStats.release() // This counts as a free, so update heapLive.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:44:56 UTC 2024 - 37.9K bytes - Viewed (0) -
src/runtime/metrics_test.go
case "/gc/heap/tiny/allocs:objects": // Currently, MemStats adds tiny alloc count to both Mallocs AND Frees. // The reason for this is because MemStats couldn't be extended at the time // but there was a desire to have Mallocs at least be a little more representative, // while having Mallocs - Frees still represent a live object count. // Unfortunately, MemStats doesn't actually export a large allocation count,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 45K bytes - Viewed (0) -
src/runtime/mgcsweep.go
// wasn't totally filled, but then swept, still has all of its // free slots zeroed. s.needzero = 1 stats := memstats.heapStats.acquire() atomic.Xadd64(&stats.smallFreeCount[spc.sizeclass()], int64(nfreed)) memstats.heapStats.release() // Count the frees in the inconsistent, internal stats. gcController.totalFree.Add(int64(nfreed) * int64(s.elemsize)) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:52:18 UTC 2024 - 32.9K bytes - Viewed (0) -
src/runtime/stack_test.go
const ( BatchSize = 32 BatchCount = 256 ArraySize = 1024 RecursionDepth = 128 ) if testing.Short() { return } defer GOMAXPROCS(GOMAXPROCS(BatchSize)) s0 := new(MemStats) ReadMemStats(s0) for b := 0; b < BatchCount; b++ { c := make(chan bool, BatchSize) for i := 0; i < BatchSize; i++ { go func() { var f func(k int, a [ArraySize]byte)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 14 00:03:57 UTC 2023 - 23.1K bytes - Viewed (0)