- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 15 for deferpool (0.54 sec)
-
src/runtime/panic.go
pp := mp.p.ptr() if len(pp.deferpool) == 0 && sched.deferpool != nil { lock(&sched.deferlock) for len(pp.deferpool) < cap(pp.deferpool)/2 && sched.deferpool != nil { d := sched.deferpool sched.deferpool = d.link d.link = nil pp.deferpool = append(pp.deferpool, d) } unlock(&sched.deferlock) } if n := len(pp.deferpool); n > 0 { d = pp.deferpool[n-1] pp.deferpool[n-1] = nil
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 43.8K bytes - Viewed (0) -
src/runtime/runtime2.go
sysmontick sysmontick // last tick observed by sysmon m muintptr // back-link to associated m (nil if idle) mcache *mcache pcache pageCache raceprocctx uintptr deferpool []*_defer // pool of available defer structs (see panic.go) deferpoolbuf [32]*_defer // Cache of goroutine ids, amortizes accesses to runtime·sched.goidgen. goidcache uint64 goidcacheend uint64
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 47.9K bytes - Viewed (0) -
src/runtime/mgc.go
// disconnect cached list before dropping it on the floor, // so that a dangling ref to one entry does not pin all of them. var d, dlink *_defer for d = sched.deferpool; d != nil; d = dlink { dlink = d.link d.link = nil } sched.deferpool = nil unlock(&sched.deferlock) } // Timing // itoaDiv formats val/(10**dec) into buf. func itoaDiv(buf []byte, val uint64, dec int) []byte {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
src/runtime/proc.go
// previously destroyed p, and transitions it to status _Pgcstop. func (pp *p) init(id int32) { pp.id = id pp.status = _Pgcstop pp.sudogcache = pp.sudogbuf[:0] pp.deferpool = pp.deferpoolbuf[:0] pp.wbBuf.reset() if pp.mcache == nil { if id == 0 { if mcache0 == nil { throw("missing mcache?") } // Use the bootstrap mcache0. Only one P will get
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 207.5K bytes - Viewed (0) -
src/cmd/compile/internal/ir/symtab.go
AssertI2I *obj.LSym AssertI2I2 *obj.LSym Asanread *obj.LSym Asanwrite *obj.LSym CgoCheckMemmove *obj.LSym CgoCheckPtrWrite *obj.LSym CheckPtrAlignment *obj.LSym Deferproc *obj.LSym Deferprocat *obj.LSym DeferprocStack *obj.LSym Deferreturn *obj.LSym Duffcopy *obj.LSym Duffzero *obj.LSym GCWriteBarrier [8]*obj.LSym
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Oct 06 17:02:26 UTC 2023 - 2K bytes - Viewed (0) -
src/cmd/internal/goobj/mkbuiltin.go
} return result } type extra struct { name string abi int } var fextras = [...]extra{ // compiler frontend inserted calls (sysfunc) {"deferproc", 1}, {"deferprocStack", 1}, {"deferreturn", 1}, {"newproc", 1}, {"panicoverflow", 1}, {"sigpanic", 1}, // compiler backend inserted calls {"gcWriteBarrier", 1}, {"duffzero", 1},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:27 UTC 2023 - 3.8K bytes - Viewed (0) -
src/cmd/internal/goobj/builtinlist.go
{"runtime.x86HasPOPCNT", 0}, {"runtime.x86HasSSE41", 0}, {"runtime.x86HasFMA", 0}, {"runtime.armHasVFPv4", 0}, {"runtime.arm64HasATOMICS", 0}, {"runtime.asanregisterglobals", 1}, {"runtime.deferproc", 1}, {"runtime.deferprocStack", 1}, {"runtime.deferreturn", 1}, {"runtime.newproc", 1}, {"runtime.panicoverflow", 1}, {"runtime.sigpanic", 1}, {"runtime.gcWriteBarrier", 1}, {"runtime.duffzero", 1},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 21:08:03 UTC 2024 - 7.4K bytes - Viewed (0) -
src/runtime/stubs.go
// See go.dev/issue/67401. // //go:linkname morestack_noctxt func morestack_noctxt() func rt0_go() // return0 is a stub used to return 0 from deferproc. // It is called at the very end of deferproc to signal // the calling Go function that it should not jump // to deferreturn. // in asm_*.s func return0() // in asm_*.s
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 20.2K bytes - Viewed (0) -
cmd/erasure-server-pool.go
return pinfo, noReadQuorumPools, nil } } if opts.ReplicationRequest && opts.DeleteMarker && defPool.Index >= 0 { // If the request is a delete marker replication request, return a default pool // in cases where the object does not exist. // This is to ensure that the delete marker is replicated to the destination. return defPool, noReadQuorumPools, nil }
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Thu May 30 11:58:12 UTC 2024 - 82.5K bytes - Viewed (0) -
src/runtime/stack.go
to be enough room to execute functions that refuse to check for stack overflow, either because they need to be adjacent to the actual caller's frame (deferproc) or because they handle the imminent stack overflow (morestack). For example, deferproc might call malloc, which does one of the above checks (without allocating a full frame), which might trigger a call to morestack. This sequence needs to fit in the bottom
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 41.1K bytes - Viewed (0)