Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 15 for deferpool (0.16 sec)

  1. src/runtime/panic.go

    	pp := mp.p.ptr()
    	if len(pp.deferpool) == 0 && sched.deferpool != nil {
    		lock(&sched.deferlock)
    		for len(pp.deferpool) < cap(pp.deferpool)/2 && sched.deferpool != nil {
    			d := sched.deferpool
    			sched.deferpool = d.link
    			d.link = nil
    			pp.deferpool = append(pp.deferpool, d)
    		}
    		unlock(&sched.deferlock)
    	}
    	if n := len(pp.deferpool); n > 0 {
    		d = pp.deferpool[n-1]
    		pp.deferpool[n-1] = nil
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 43.8K bytes
    - Viewed (0)
  2. src/runtime/runtime2.go

    	sysmontick  sysmontick // last tick observed by sysmon
    	m           muintptr   // back-link to associated m (nil if idle)
    	mcache      *mcache
    	pcache      pageCache
    	raceprocctx uintptr
    
    	deferpool    []*_defer // pool of available defer structs (see panic.go)
    	deferpoolbuf [32]*_defer
    
    	// Cache of goroutine ids, amortizes accesses to runtime·sched.goidgen.
    	goidcache    uint64
    	goidcacheend uint64
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 47.9K bytes
    - Viewed (0)
  3. src/runtime/mgc.go

    	// disconnect cached list before dropping it on the floor,
    	// so that a dangling ref to one entry does not pin all of them.
    	var d, dlink *_defer
    	for d = sched.deferpool; d != nil; d = dlink {
    		dlink = d.link
    		d.link = nil
    	}
    	sched.deferpool = nil
    	unlock(&sched.deferlock)
    }
    
    // Timing
    
    // itoaDiv formats val/(10**dec) into buf.
    func itoaDiv(buf []byte, val uint64, dec int) []byte {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 62K bytes
    - Viewed (0)
  4. src/runtime/proc.go

    // previously destroyed p, and transitions it to status _Pgcstop.
    func (pp *p) init(id int32) {
    	pp.id = id
    	pp.status = _Pgcstop
    	pp.sudogcache = pp.sudogbuf[:0]
    	pp.deferpool = pp.deferpoolbuf[:0]
    	pp.wbBuf.reset()
    	if pp.mcache == nil {
    		if id == 0 {
    			if mcache0 == nil {
    				throw("missing mcache?")
    			}
    			// Use the bootstrap mcache0. Only one P will get
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 207.5K bytes
    - Viewed (0)
  5. src/cmd/compile/internal/ir/symtab.go

    	AssertI2I         *obj.LSym
    	AssertI2I2        *obj.LSym
    	Asanread          *obj.LSym
    	Asanwrite         *obj.LSym
    	CgoCheckMemmove   *obj.LSym
    	CgoCheckPtrWrite  *obj.LSym
    	CheckPtrAlignment *obj.LSym
    	Deferproc         *obj.LSym
    	Deferprocat       *obj.LSym
    	DeferprocStack    *obj.LSym
    	Deferreturn       *obj.LSym
    	Duffcopy          *obj.LSym
    	Duffzero          *obj.LSym
    	GCWriteBarrier    [8]*obj.LSym
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Oct 06 17:02:26 UTC 2023
    - 2K bytes
    - Viewed (0)
  6. src/cmd/internal/goobj/mkbuiltin.go

    	}
    	return result
    }
    
    type extra struct {
    	name string
    	abi  int
    }
    
    var fextras = [...]extra{
    	// compiler frontend inserted calls (sysfunc)
    	{"deferproc", 1},
    	{"deferprocStack", 1},
    	{"deferreturn", 1},
    	{"newproc", 1},
    	{"panicoverflow", 1},
    	{"sigpanic", 1},
    
    	// compiler backend inserted calls
    	{"gcWriteBarrier", 1},
    	{"duffzero", 1},
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Oct 19 23:33:27 UTC 2023
    - 3.8K bytes
    - Viewed (0)
  7. src/cmd/internal/goobj/builtinlist.go

    	{"runtime.x86HasPOPCNT", 0},
    	{"runtime.x86HasSSE41", 0},
    	{"runtime.x86HasFMA", 0},
    	{"runtime.armHasVFPv4", 0},
    	{"runtime.arm64HasATOMICS", 0},
    	{"runtime.asanregisterglobals", 1},
    	{"runtime.deferproc", 1},
    	{"runtime.deferprocStack", 1},
    	{"runtime.deferreturn", 1},
    	{"runtime.newproc", 1},
    	{"runtime.panicoverflow", 1},
    	{"runtime.sigpanic", 1},
    	{"runtime.gcWriteBarrier", 1},
    	{"runtime.duffzero", 1},
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue May 21 21:08:03 UTC 2024
    - 7.4K bytes
    - Viewed (0)
  8. src/runtime/stubs.go

    // See go.dev/issue/67401.
    //
    //go:linkname morestack_noctxt
    func morestack_noctxt()
    
    func rt0_go()
    
    // return0 is a stub used to return 0 from deferproc.
    // It is called at the very end of deferproc to signal
    // the calling Go function that it should not jump
    // to deferreturn.
    // in asm_*.s
    func return0()
    
    // in asm_*.s
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 20.2K bytes
    - Viewed (0)
  9. cmd/erasure-server-pool.go

    			return pinfo, noReadQuorumPools, nil
    		}
    	}
    	if opts.ReplicationRequest && opts.DeleteMarker && defPool.Index >= 0 {
    		// If the request is a delete marker replication request, return a default pool
    		// in cases where the object does not exist.
    		// This is to ensure that the delete marker is replicated to the destination.
    		return defPool, noReadQuorumPools, nil
    	}
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Thu May 30 11:58:12 UTC 2024
    - 82.5K bytes
    - Viewed (0)
  10. src/runtime/stack.go

    to be enough room to execute functions that refuse to check for
    stack overflow, either because they need to be adjacent to the
    actual caller's frame (deferproc) or because they handle the imminent
    stack overflow (morestack).
    
    For example, deferproc might call malloc, which does one of the
    above checks (without allocating a full frame), which might trigger
    a call to morestack.  This sequence needs to fit in the bottom
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 41.1K bytes
    - Viewed (0)
Back to top