- Sort Score
- Result 10 results
- Languages All
Results 61 - 70 of 81 for pcs (0.03 sec)
-
src/debug/dwarf/entry.go
return nil, err } if e == nil || e.Tag == 0 { return nil, ErrUnknownPC } ranges, err := r.d.Ranges(e) if err != nil { return nil, err } for _, pcs := range ranges { if pcs[0] <= pc && pc < pcs[1] { return e, nil } } unit++ } return nil, ErrUnknownPC } // Ranges returns the PC ranges covered by e, a slice of [low,high) pairs.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Oct 18 19:33:30 UTC 2023 - 30.7K bytes - Viewed (0) -
src/internal/bisect/bisect.go
func (m *Matcher) stack(w Writer) bool { const maxStack = 16 var stk [maxStack]uintptr n := runtime.Callers(2, stk[:]) // caller #2 is not for printing; need it to normalize PCs if ASLR. if n <= 1 { return false } base := stk[0] // normalize PCs for i := range stk[:n] { stk[i] -= base } h := Hash(stk[:n]) if m.ShouldPrint(h) { var d *dedup for { d = m.dedup.Load()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 17:28:43 UTC 2024 - 22.9K bytes - Viewed (0) -
src/runtime/traceevent.go
// stack ID for it. // // It then returns a traceArg representing that stack which may be // passed to write. func (tl traceLocker) startPC(pc uintptr) traceArg { // +PCQuantum because makeTraceFrame expects return PCs and subtracts PCQuantum. return traceArg(trace.stackTab[tl.gen%2].put([]uintptr{ logicalStackSentinel, startPCForTrace(pc) + sys.PCQuantum, })) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:47:01 UTC 2024 - 9.2K bytes - Viewed (0) -
src/cmd/link/internal/ld/dwarf_test.go
// - at least one row has the correct line number (8) pcs := make(map[uint64]bool) line8seen := false for _, r := range rows { pcs[r.Address] = true if r.Line == 8 { line8seen = true } } failed := false if len(pcs) < 2 { failed = true t.Errorf("not enough line table rows for main.singleInstruction (got %d, wanted > 1", len(pcs)) } if !line8seen { failed = true
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Feb 08 01:38:11 UTC 2024 - 48.6K bytes - Viewed (0) -
src/debug/dwarf/line.go
// basic block. BasicBlock bool // PrologueEnd indicates that Address is one (of possibly // many) PCs where execution should be suspended for a // breakpoint on entry to the containing function. // // Added in DWARF 3. PrologueEnd bool // EpilogueBegin indicates that Address is one (of possibly // many) PCs where execution should be suspended for a // breakpoint on exit from this function. // // Added in DWARF 3.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Oct 18 19:33:30 UTC 2023 - 23.5K bytes - Viewed (0) -
src/io/multi_test.go
type readerFunc func(p []byte) (int, error) func (f readerFunc) Read(p []byte) (int, error) { return f(p) } // callDepth returns the logical call depth for the given PCs. func callDepth(callers []uintptr) (depth int) { frames := runtime.CallersFrames(callers) more := true for more { _, more = frames.Next() depth++ } return }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Sep 06 15:49:32 UTC 2022 - 10K bytes - Viewed (0) -
src/runtime/pprof/pprof.go
expandedStack := pprof_makeProfStack() for _, r := range records { values[0] = r.Count values[1] = int64(float64(r.Cycles) / cpuGHz) // For count profiles, all stack addresses are // return PCs, which is what appendLocsForStack expects. n := pprof_fpunwindExpand(expandedStack[:], r.Stack) locs = b.appendLocsForStack(locs[:0], expandedStack[:n]) b.pbSample(values, locs, nil) } b.build() return nil
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 17:52:17 UTC 2024 - 30.6K bytes - Viewed (0) -
src/internal/trace/event.go
// Frames is an iterator over the frames in a Stack. func (s Stack) Frames(yield func(f StackFrame) bool) bool { if s.id == 0 { return true } stk := s.table.stacks.mustGet(s.id) for _, pc := range stk.pcs { f := s.table.pcs[pc] sf := StackFrame{ PC: f.pc, Func: s.table.strings.mustGet(f.funcID), File: s.table.strings.mustGet(f.fileID), Line: f.line, } if !yield(sf) { return false }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 30 12:39:00 UTC 2024 - 28.9K bytes - Viewed (0) -
src/runtime/preempt.go
func init() { f := findfunc(abi.FuncPCABI0(asyncPreempt)) total := funcMaxSPDelta(f) f = findfunc(abi.FuncPCABIInternal(asyncPreempt2)) total += funcMaxSPDelta(f) // Add some overhead for return PCs, etc. asyncPreemptStack = uintptr(total) + 8*goarch.PtrSize if asyncPreemptStack > stackNosplit { // We need more than the nosplit limit. This isn't // unsafe, but it may limit asynchronous preemption. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 15:41:45 UTC 2024 - 15.1K bytes - Viewed (0) -
src/cmd/internal/obj/pcln.go
} pcln := arg.(*Pcln) pcln.UsedFiles[goobj.CUFileIndex(f)] = struct{}{} return int32(f) } // pcinlineState holds the state used to create a function's inlining // tree and the PC-value table that maps PCs to nodes in that tree. type pcinlineState struct { globalToLocal map[int]int localTree InlTree } // addBranch adds a branch from the global inlining tree in ctxt to
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Oct 31 20:45:15 UTC 2022 - 11.8K bytes - Viewed (0)