Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 3 of 3 for profStack (0.17 sec)

  1. src/runtime/mprof.go

    				// saveblockevent)
    				mp.profStack[0] -= 1
    			}
    			nstk += fpTracebackPCs(unsafe.Pointer(getfp()), mp.profStack[1:])
    		} else {
    			mp.profStack[1] = gp.m.curg.sched.pc
    			nstk += 1 + fpTracebackPCs(unsafe.Pointer(gp.m.curg.sched.bp), mp.profStack[2:])
    		}
    	}
    
    	saveBlockEventStack(cycles, rate, mp.profStack[:nstk], which)
    	releasem(mp)
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 53.3K bytes
    - Viewed (0)
  2. src/runtime/runtime2.go

    	lockedInt     uint32      // tracking for internal lockOSThread
    	nextwaitm     muintptr    // next m waiting for lock
    
    	mLockProfile mLockProfile // fields relating to runtime.lock contention
    	profStack    []uintptr    // used for memory/block/mutex stack traces
    
    	// wait* are used to carry arguments from gopark into park_m, because
    	// there's no stack to put them on. That is their sole purpose.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 30 17:57:37 UTC 2024
    - 47.9K bytes
    - Viewed (0)
  3. src/runtime/proc.go

    // TODO(mknyszek): Implement lazy allocation if this becomes a problem.
    func mProfStackInit(mp *m) {
    	if debug.profstackdepth == 0 {
    		// debug.profstack is set to 0 by the user, or we're being called from
    		// schedinit before parsedebugvars.
    		return
    	}
    	mp.profStack = makeProfStackFP()
    	mp.mLockProfile.stack = makeProfStackFP()
    }
    
    // makeProfStackFP creates a buffer large enough to hold a maximum-sized stack
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 207.5K bytes
    - Viewed (0)
Back to top