- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 694 for Atack (0.02 sec)
-
src/runtime/signal_mips64x.go
// preparePanic sets up the stack to look like a call to sigpanic. func (c *sigctxt) preparePanic(sig uint32, gp *g) { // We arrange link, and pc to pretend the panicking // function calls sigpanic directly. // Always save LINK to stack so that panics in leaf // functions are correctly handled. This smashes // the stack frame but we're not going back there // anyway. sp := c.sp() - goarch.PtrSize
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 28 18:17:57 UTC 2021 - 3.2K bytes - Viewed (0) -
test/fixedbugs/issue16016.go
_ = v } } type R struct{ *T } type Q interface { Foo([]interface{}) } func main() { var count = 10000 if runtime.Compiler == "gccgo" { // On targets without split-stack libgo allocates // a large stack for each goroutine. On 32-bit // systems this test can run out of memory. const intSize = 32 << (^uint(0) >> 63) // 32 or 64 if intSize < 64 { count = 100 } } var q Q = &R{&T{}}
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 27 18:39:06 UTC 2024 - 877 bytes - Viewed (0) -
src/runtime/mem_windows.go
_PAGE_READWRITE = 0x0004 _PAGE_NOACCESS = 0x0001 _ERROR_NOT_ENOUGH_MEMORY = 8 _ERROR_COMMITMENT_LIMIT = 1455 ) // Don't split the stack as this function may be invoked without a valid G, // which prevents us from allocating more stack. // //go:nosplit func sysAllocOS(n uintptr) unsafe.Pointer { return unsafe.Pointer(stdcall4(_VirtualAlloc, 0, n, _MEM_COMMIT|_MEM_RESERVE, _PAGE_READWRITE)) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 3.9K bytes - Viewed (0) -
test/fixedbugs/issue16515.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jul 29 01:09:55 UTC 2016 - 1.1K bytes - Viewed (0) -
src/cmd/compile/internal/test/mergelocals_test.go
testenv.MustHaveGoBuild(t) // This test does a build of a specific canned package to // check whether merging of stack slots is taking place. // The idea is to do the compile with a trace option turned // on and then pick up on the frame offsets of specific // variables. // // Stack slot merging is a greedy algorithm, and there can // be many possible ways to overlap a given set of candidate
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 15:43:53 UTC 2024 - 4.7K bytes - Viewed (0) -
src/runtime/mstats.go
// freed. HeapObjects uint64 // Stack memory statistics. // // Stacks are not considered part of the heap, but the runtime // can reuse a span of heap memory for stack memory, and // vice-versa. // StackInuse is bytes in stack spans. // // In-use stack spans have at least one stack in them. These // spans can only be used for other stacks of the same size. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 08 21:03:13 UTC 2024 - 34.2K bytes - Viewed (0) -
test/fixedbugs/issue29919.dir/a.go
package a import ( "fmt" "runtime" "strings" ) var x = f() // line 15 func f() int { var b [4096]byte n := runtime.Stack(b[:], false) // line 19 s := string(b[:n]) var pcs [10]uintptr n = runtime.Callers(1, pcs[:]) // line 22 // Check the Stack results. if debug { println(s) } if strings.Contains(s, "autogenerated") { panic("autogenerated code in traceback") }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Mar 24 17:50:47 UTC 2022 - 1.6K bytes - Viewed (0) -
src/runtime/testdata/testprog/traceback_ancestors.go
recurseThenCallGo(w, numGoroutines, numFrames, true) <-w printStack() close(w) } var ignoreGoroutines = make(map[string]bool) func printStack() { buf := make([]byte, 1024) for { n := runtime.Stack(buf, true) if n < len(buf) { all := string(buf[:n]) var saved string // Delete any ignored goroutines, if present. for all != "" { var g string g, all, _ = strings.Cut(all, "\n\n")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Sep 27 17:14:59 UTC 2022 - 2.1K bytes - Viewed (0) -
src/runtime/mgc.go
} systemstack(func() { // Mark our goroutine preemptible so its stack // can be scanned. This lets two mark workers // scan each other (otherwise, they would // deadlock). We must not modify anything on // the G stack. However, stack shrinking is // disabled for mark workers, so it is safe to // read from the G stack. // // N.B. The execution tracer is not aware of this status
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 62K bytes - Viewed (0) -
src/runtime/mwbbuf.go
// the buffer if the stack has been shaded, or even avoid // putting them in the buffer at all (which would double its // capacity). This is slightly complicated with the buffer; we // could track whether any un-shaded goroutine has used the // buffer, or just track globally whether there are any // un-shaded stacks and flush after each stack scan. gcw := &pp.gcw pos := 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0)