Search Options

Results per page
Sort
Preferred Languages
Advance

Results 71 - 80 of 83 for sched (0.04 sec)

  1. src/runtime/stubs.go

    func getg() *g
    
    // mcall switches from the g to the g0 stack and invokes fn(g),
    // where g is the goroutine that made the call.
    // mcall saves g's current PC/SP in g->sched so that it can be restored later.
    // It is up to fn to arrange for that later execution, typically by recording
    // g in a data structure, causing something to call ready(g) later.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 20.2K bytes
    - Viewed (0)
  2. src/runtime/extern.go

    		# MB stacks  estimated scannable stack size, or /gc/scan/stack:bytes
    		# MB globals scannable global size, or /gc/scan/globals:bytes
    		# P          number of processors used, or /sched/gomaxprocs:threads
    	The phases are stop-the-world (STW) sweep termination, concurrent
    	mark and scan, and STW mark termination. The CPU times
    	for mark/scan are broken down in to assist time (GC performed in
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:52:17 UTC 2024
    - 18.9K bytes
    - Viewed (0)
  3. src/runtime/mprof.go

    				// saveblockevent)
    				mp.profStack[0] -= 1
    			}
    			nstk += fpTracebackPCs(unsafe.Pointer(getfp()), mp.profStack[1:])
    		} else {
    			mp.profStack[1] = gp.m.curg.sched.pc
    			nstk += 1 + fpTracebackPCs(unsafe.Pointer(gp.m.curg.sched.bp), mp.profStack[2:])
    		}
    	}
    
    	saveBlockEventStack(cycles, rate, mp.profStack[:nstk], which)
    	releasem(mp)
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 53.3K bytes
    - Viewed (0)
  4. src/internal/trace/gc.go

    	for i := range events {
    		ev := &events[i]
    		lastEv = ev
    
    		// Process the event.
    		switch ev.Kind() {
    		case EventSync:
    			seenSync = true
    		case EventMetric:
    			m := ev.Metric()
    			if m.Name != "/sched/gomaxprocs:threads" {
    				break
    			}
    			gomaxprocs := int(m.Value.Uint64())
    			if len(ps) > gomaxprocs {
    				if flags&UtilPerProc != 0 {
    					// End each P's series.
    					for _, p := range ps[gomaxprocs:] {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 18:48:18 UTC 2024
    - 26K bytes
    - Viewed (0)
  5. src/runtime/mgcpacer.go

    //go:nowritebarrier
    func (c *gcControllerState) enlistWorker() {
    	// If there are idle Ps, wake one so it will run an idle worker.
    	// NOTE: This is suspected of causing deadlocks. See golang.org/issue/19112.
    	//
    	//	if sched.npidle.Load() != 0 && sched.nmspinning.Load() == 0 {
    	//		wakep()
    	//		return
    	//	}
    
    	// There are no idle Ps. If we need more dedicated workers,
    	// try to preempt a running P so it will switch to a worker.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 25 19:53:03 UTC 2024
    - 55.4K bytes
    - Viewed (0)
  6. src/internal/trace/event.go

    func (e Event) Metric() Metric {
    	if e.Kind() != EventMetric {
    		panic("Metric called on non-Metric event")
    	}
    	var m Metric
    	switch e.base.typ {
    	case go122.EvProcsChange:
    		m.Name = "/sched/gomaxprocs:threads"
    		m.Value = Value{kind: ValueUint64, scalar: e.base.args[0]}
    	case go122.EvHeapAlloc:
    		m.Name = "/memory/classes/heap/objects:bytes"
    		m.Value = Value{kind: ValueUint64, scalar: e.base.args[0]}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 12:39:00 UTC 2024
    - 28.9K bytes
    - Viewed (0)
  7. src/runtime/mbitmap.go

    		KeepAlive(ep)
    		return
    	}
    
    	// stack
    	if gp := getg(); gp.m.curg.stack.lo <= uintptr(p) && uintptr(p) < gp.m.curg.stack.hi {
    		found := false
    		var u unwinder
    		for u.initAt(gp.m.curg.sched.pc, gp.m.curg.sched.sp, 0, gp.m.curg, 0); u.valid(); u.next() {
    			if u.frame.sp <= uintptr(p) && uintptr(p) < u.frame.varp {
    				found = true
    				break
    			}
    		}
    		if found {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 00:18:55 UTC 2024
    - 60K bytes
    - Viewed (0)
  8. src/runtime/time.go

    func (ts *timers) take(src *timers) {
    	ts.trace("take")
    	assertWorldStopped()
    	if len(src.heap) > 0 {
    		// The world is stopped, so we ignore the locking of ts and src here.
    		// That would introduce a sched < timers lock ordering,
    		// which we'd rather avoid in the static ranking.
    		for _, tw := range src.heap {
    			t := tw.timer
    			t.ts = nil
    			if t.state&timerZombie != 0 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Mar 29 14:36:24 UTC 2024
    - 37.5K bytes
    - Viewed (0)
  9. src/internal/trace/order.go

    			if ms.p == pid {
    				curCtx.M = mid
    				curCtx.P = pid
    				curCtx.G = ms.g
    				found = true
    			}
    		}
    		if !found {
    			return curCtx, false, fmt.Errorf("failed to find sched context for proc %d that's about to be stolen", pid)
    		}
    	}
    	o.queue.push(Event{table: evt, ctx: curCtx, base: *ev})
    	return newCtx, true, nil
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Jun 03 14:56:25 UTC 2024
    - 52.4K bytes
    - Viewed (0)
  10. src/expvar/expvar_test.go

    	var s String
    
    	b.RunParallel(func(pb *testing.PB) {
    		for pb.Next() {
    			s.Set("red")
    		}
    	})
    }
    
    func TestMapInit(t *testing.T) {
    	RemoveAll()
    	colors := NewMap("bike-shed-colors")
    	colors.Add("red", 1)
    	colors.Add("blue", 1)
    	colors.Add("chartreuse", 1)
    
    	n := 0
    	colors.Do(func(KeyValue) { n++ })
    	if n != 3 {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Apr 04 14:46:19 UTC 2024
    - 13.4K bytes
    - Viewed (0)
Back to top