- Sort Score
- Result 10 results
- Languages All
Results 71 - 80 of 174 for Atack (0.07 sec)
-
src/runtime/mem_bsd.go
//go:build dragonfly || freebsd || netbsd || openbsd || solaris package runtime import ( "unsafe" ) // Don't split the stack as this function may be invoked without a valid G, // which prevents us from allocating more stack. // //go:nosplit func sysAllocOS(n uintptr) unsafe.Pointer { v, err := mmap(nil, n, _PROT_READ|_PROT_WRITE, _MAP_ANON|_MAP_PRIVATE, -1, 0) if err != 0 { return nil
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 2.2K bytes - Viewed (0) -
src/reflect/export_test.go
} // Extract size information. argSize = abid.stackCallArgsSize retOffset = abid.retOffset frametype = toType(ft) // Expand stack pointer bitmap into byte-map. for i := uint32(0); i < abid.stackPtrs.n; i++ { stack = append(stack, abid.stackPtrs.data[i/8]>>(i%8)&1) } // Expand register pointer bitmaps into byte-maps. bool2byte := func(b bool) byte { if b { return 1 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 02 15:10:48 UTC 2024 - 3.8K bytes - Viewed (0) -
src/internal/trace/event/event.go
Args []string // StringIDs indicates which of the arguments are string IDs. StringIDs []int // StackIDs indicates which of the arguments are stack IDs. // // The list is not sorted. The first index always refers to // the main stack for the current execution context of the event. StackIDs []int // StartEv indicates the event type of the corresponding "start"
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 3.4K bytes - Viewed (0) -
src/internal/trace/testdata/testprog/cpu-profile.go
pprofStacks[stack] += samples } } for stack, samples := range pprofStacks { fmt.Fprintf(os.Stderr, "%s\t%d\n", stack, samples) } } func cpuHogger(f func(x int) int, y *int, dur time.Duration) { // We only need to get one 100 Hz clock tick, so we've got // a large safety buffer. // But do at least 500 iterations (which should take about 100ms), // otherwise TestCPUProfileMultithreaded can fail if only one
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 3.8K bytes - Viewed (0) -
src/runtime/debug/stack_test.go
func (t *T) ptrmethod() []byte { return Stack() } func (t T) method() []byte { return t.ptrmethod() } /* The traceback should look something like this, modulo line numbers and hex constants. Don't worry much about the base levels, but check the ones in our own package. goroutine 10 [running]: runtime/debug.Stack(0x0, 0x0, 0x0) /Users/r/go/src/runtime/debug/stack.go:28 +0x80
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 16 15:19:04 UTC 2024 - 5.5K bytes - Viewed (0) -
src/runtime/testdata/testprogcgo/numgoroutine.go
if n != want { fmt.Printf("%s NumGoroutine: want %d; got %d\n", label, want, n) return "", false } sbuf := make([]byte, 32<<10) sbuf = sbuf[:runtime.Stack(sbuf, true)] n = strings.Count(string(sbuf), "goroutine ") if n != want { fmt.Printf("%s Stack: want %d; got %d:\n%s\n", label, want, n, sbuf) return "", false } return string(sbuf), true } var callbackok bool //export CallbackNumGoroutine
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 22:52:37 UTC 2023 - 1.9K bytes - Viewed (0) -
src/go/types/mono.go
// vertices we visited earlier cannot be part of the cycle. for stack[0] != v { stack = stack[1:] } // TODO(mdempsky): Pivot stack so we report the cycle from the top? err := check.newError(InvalidInstanceCycle) obj0 := check.mono.vertices[v].obj err.addf(obj0, "instantiation cycle:") qf := RelativeTo(check.pkg) for _, v := range stack { edge := check.mono.edges[check.mono.vertices[v].pre]
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Apr 03 18:48:38 UTC 2024 - 9.2K bytes - Viewed (0) -
src/cmd/trace/viewer.go
package main import ( "fmt" "internal/trace" "internal/trace/traceviewer" "time" ) // viewerFrames returns the frames of the stack of ev. The given frame slice is // used to store the frames to reduce allocations. func viewerFrames(stk trace.Stack) []*trace.Frame { var frames []*trace.Frame stk.Frames(func(f trace.StackFrame) bool { frames = append(frames, &trace.Frame{ PC: f.PC, Fn: f.Func,
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:48:18 UTC 2024 - 1.4K bytes - Viewed (0) -
src/runtime/mem.go
// or a megabyte. This memory is always immediately available for use. // // sysStat must be non-nil. // // Don't split the stack as this function may be invoked without a valid G, // which prevents us from allocating more stack. // //go:nosplit func sysAlloc(n uintptr, sysStat *sysMemStat) unsafe.Pointer { sysStat.add(int64(n)) gcController.mappedReady.Add(int64(n)) return sysAllocOS(n) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 6.7K bytes - Viewed (0) -
test/fixedbugs/issue42284.dir/b.go
package b import "./a" func g() { h := a.E() // ERROR "inlining call to a.E" "T\(0\) does not escape" h.M() // ERROR "devirtualizing h.M to a.T" "inlining call to a.T.M" // BAD: T(0) could be stack allocated. i := a.F(a.T(0)) // ERROR "inlining call to a.F" "a.T\(0\) escapes to heap" // Testing that we do NOT devirtualize here: i.M()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Nov 20 18:09:45 UTC 2023 - 513 bytes - Viewed (0)