- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 74 for systemstack (0.3 sec)
-
src/runtime/pprof/pprof_test.go
for i := 0; i < 10; i++ { goroutineProf.WriteTo(io.Discard, 1) } }) } // TestLabelSystemstack makes sure CPU profiler samples of goroutines running // on systemstack include the correct pprof labels. See issue #48577 func TestLabelSystemstack(t *testing.T) { // Grab and re-set the initial value before continuing to ensure // GOGC doesn't actually change following the test.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 18:42:28 UTC 2024 - 68.8K bytes - Viewed (0) -
src/runtime/traceruntime.go
// synchronize with the tracer trying to flush our buffer // as well. seq := mp.trace.seqlock.Add(1) if debugTraceReentrancy && seq%2 != 1 { throw("bad use of trace.seqlock or tracer is reentrant") } systemstack(func() { lock(&trace.lock) for i := range mp.trace.buf { if mp.trace.buf[i] != nil { // N.B. traceBufFlush accepts a generation, but it // really just cares about gen%2.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 25.7K bytes - Viewed (0) -
src/runtime/cgocall.go
gp.m.winsyscall = winsyscall } func cgocallbackg1(fn, frame unsafe.Pointer, ctxt uintptr) { gp := getg() if gp.m.needextram || extraMWaiters.Load() > 0 { gp.m.needextram = false systemstack(newextram) } if ctxt != 0 { s := append(gp.cgoCtxt, ctxt) // Now we need to set gp.cgoCtxt = s, but we could get // a SIGPROF signal while manipulating the slice, and
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:16:47 UTC 2024 - 24.2K bytes - Viewed (0) -
src/runtime/mgcpacer.go
c.gcPercent.Store(in) return out } //go:linkname setGCPercent runtime/debug.setGCPercent func setGCPercent(in int32) (out int32) { // Run on the system stack since we grab the heap lock. systemstack(func() { lock(&mheap_.lock) out = gcController.setGCPercent(in) gcControllerCommit() unlock(&mheap_.lock) }) // If we just disabled GC, wait for any concurrent GC mark to
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 55.4K bytes - Viewed (0) -
src/runtime/mgcscavenge.go
func (p *pageAlloc) scavenge(nbytes uintptr, shouldStop func() bool, force bool) uintptr { released := uintptr(0) for released < nbytes { ci, pageIdx := p.scav.index.find(force) if ci == 0 { break } systemstack(func() { released += p.scavengeOne(ci, pageIdx, nbytes-released) }) if shouldStop != nil && shouldStop() { break } } return released }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:48:45 UTC 2024 - 52.3K bytes - Viewed (0) -
src/runtime/signal_unix.go
// and printable name of each signal. type sigTabT struct { flags int32 name string } //go:linkname os_sigpipe os.sigpipe func os_sigpipe() { systemstack(sigpipe) } func signame(sig uint32) string { if sig >= uint32(len(sigtable)) { return "" } return sigtable[sig].name } const ( _SIG_DFL uintptr = 0 _SIG_IGN uintptr = 1 )
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 10 16:04:54 UTC 2024 - 45K bytes - Viewed (0) -
src/runtime/metrics_test.go
acceptStacks = append([][]string(nil), acceptStacks...) for i, stk := range acceptStacks { if goexperiment.StaticLockRanking { if !slices.ContainsFunc(stk, func(s string) bool { return s == "runtime.systemstack" || s == "runtime.mcall" || s == "runtime.mstart" }) { // stk is a call stack that is still on the user stack when // it calls runtime.unlock. Add the extra function that
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 45K bytes - Viewed (0) -
src/runtime/malloc.go
// // Consider marking persistentalloc'd types not in heap by embedding // runtime/internal/sys.NotInHeap. func persistentalloc(size, align uintptr, sysStat *sysMemStat) unsafe.Pointer { var p *notInHeap systemstack(func() { p = persistentalloc1(size, align, sysStat) }) return unsafe.Pointer(p) } // Must run on system stack because stack growth can (re)invoke it. // See issue 9174. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 59.6K bytes - Viewed (0) -
src/runtime/mbitmap.go
heapBitsOff := spaceNeeded spaceNeeded += alignUp(typ.PtrBytes/goarch.PtrSize/8, goarch.PtrSize) npages := alignUp(spaceNeeded, pageSize) / pageSize var progSpan *mspan systemstack(func() { progSpan = mheap_.allocManual(npages, spanAllocPtrScalarBits) memclrNoHeapPointers(unsafe.Pointer(progSpan.base()), progSpan.npages*pageSize) }) // Write a dummy _type in the new space. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
LICENSES/vendor/bitbucket.org/bertimus9/systemstat/LICENSE
= vendor/bitbucket.org/bertimus9/systemstat licensed under: = The MIT License (MIT) Copyright (c) 2013 Phillip Bond Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Fri May 08 04:49:00 UTC 2020 - 1.2K bytes - Viewed (0)