- Sort Score
- Result 10 results
- Languages All
Results 1 - 5 of 5 for tracefpunwindoff (0.29 sec)
-
src/runtime/trace_cgo_test.go
"goCalledFromC", "goCalledFromCThread", } logs := make(map[string]*trace.Event) for _, category := range wantLogs { logs[category] = nil } for _, tracefpunwindoff := range []int{1, 0} { env := fmt.Sprintf("GODEBUG=tracefpunwindoff=%d", tracefpunwindoff) got := runBuiltTestProg(t, exe, "Trace", env) prefix, tracePath, found := strings.Cut(got, ":") if !found || prefix != "trace path" {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 2.6K bytes - Viewed (0) -
src/runtime/tracestack.go
file = file[len(file)-maxLen:] } frame.fileID = trace.stringTab[gen%2].put(gen, file) return frame } // tracefpunwindoff returns true if frame pointer unwinding for the tracer is // disabled via GODEBUG or not supported by the architecture. func tracefpunwindoff() bool { return debug.tracefpunwindoff != 0 || (goarch.ArchFamily != goarch.AMD64 && goarch.ArchFamily != goarch.ARM64) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue May 21 14:38:56 UTC 2024 - 11K bytes - Viewed (0) -
src/runtime/runtime1.go
{name: "traceallocfree", atomic: &debug.traceallocfree}, {name: "tracecheckstackownership", value: &debug.traceCheckStackOwnership}, {name: "tracebackancestors", value: &debug.tracebackancestors}, {name: "tracefpunwindoff", value: &debug.tracefpunwindoff}, } func parsedebugvars() { // defaults debug.cgocheck = 1 debug.invalidptr = 1 debug.adaptivestackstart = 1 // set this to 0 to turn larger initial goroutine stacks off
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 19.3K bytes - Viewed (0) -
src/runtime/extern.go
which goroutines were created, where N limits the number of ancestor goroutines to report. This also extends the information returned by runtime.Stack. Setting N to 0 will report no ancestry information. tracefpunwindoff: setting tracefpunwindoff=1 forces the execution tracer to use the runtime's default stack unwinder instead of frame pointer unwinding. This increases tracer overhead, but could be helpful as a workaround or for
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 18.9K bytes - Viewed (0) -
src/runtime/mprof.go
} if skip > maxSkip { print("requested skip=", skip) throw("invalid skip value") } gp := getg() mp := acquirem() // we must not be preempted while accessing profstack nstk := 1 if tracefpunwindoff() || gp.m.hasCgoOnStack() { mp.profStack[0] = logicalStackSentinel if gp.m.curg == nil || gp.m.curg == gp { nstk = callers(skip, mp.profStack[1:]) } else {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:57:37 UTC 2024 - 53.3K bytes - Viewed (0)