- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 175 for Atack (0.12 sec)
-
test/fixedbugs/issue16016.go
_ = v } } type R struct{ *T } type Q interface { Foo([]interface{}) } func main() { var count = 10000 if runtime.Compiler == "gccgo" { // On targets without split-stack libgo allocates // a large stack for each goroutine. On 32-bit // systems this test can run out of memory. const intSize = 32 << (^uint(0) >> 63) // 32 or 64 if intSize < 64 { count = 100 } } var q Q = &R{&T{}}
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 27 18:39:06 UTC 2024 - 877 bytes - Viewed (0) -
src/runtime/mem_windows.go
_PAGE_READWRITE = 0x0004 _PAGE_NOACCESS = 0x0001 _ERROR_NOT_ENOUGH_MEMORY = 8 _ERROR_COMMITMENT_LIMIT = 1455 ) // Don't split the stack as this function may be invoked without a valid G, // which prevents us from allocating more stack. // //go:nosplit func sysAllocOS(n uintptr) unsafe.Pointer { return unsafe.Pointer(stdcall4(_VirtualAlloc, 0, n, _MEM_COMMIT|_MEM_RESERVE, _PAGE_READWRITE)) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Aug 22 19:05:10 UTC 2023 - 3.9K bytes - Viewed (0) -
src/cmd/compile/internal/test/mergelocals_test.go
testenv.MustHaveGoBuild(t) // This test does a build of a specific canned package to // check whether merging of stack slots is taking place. // The idea is to do the compile with a trace option turned // on and then pick up on the frame offsets of specific // variables. // // Stack slot merging is a greedy algorithm, and there can // be many possible ways to overlap a given set of candidate
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 15:43:53 UTC 2024 - 4.7K bytes - Viewed (0) -
src/runtime/mwbbuf.go
// the buffer if the stack has been shaded, or even avoid // putting them in the buffer at all (which would double its // capacity). This is slightly complicated with the buffer; we // could track whether any un-shaded goroutine has used the // buffer, or just track globally whether there are any // un-shaded stacks and flush after each stack scan. gcw := &pp.gcw pos := 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/runtime/mklockrank.go
profMemActive < profMemFuture; # Stack allocation and copying gcBitsArenas, netpollInit, profBlock, profInsert, profMemFuture, spanSetSpine, fin, root # Anything that can grow the stack can acquire STACKGROW. # (Most higher layers imply STACKGROW, like MALLOC.) < STACKGROW # Below STACKGROW is the stack allocator/copying implementation. < gscan; gscan < stackpool;
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:47:01 UTC 2024 - 9.1K bytes - Viewed (0) -
src/internal/trace/traceviewer/pprof.go
return } defer os.Remove(svgFilename) w.Header().Set("Content-Type", "image/svg+xml") http.ServeFile(w, r, svgFilename) } } type ProfileRecord struct { Stack []*trace.Frame Count uint64 Time time.Duration } func BuildProfile(prof []ProfileRecord) *profile.Profile { p := &profile.Profile{ PeriodType: &profile.ValueType{Type: "trace", Unit: "count"},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 21:28:02 UTC 2023 - 4K bytes - Viewed (0) -
src/regexp/backtrack.go
// // backtrack is a fast replacement for the NFA code on small // regexps when onepass cannot be used. package regexp import ( "regexp/syntax" "sync" ) // A job is an entry on the backtracker's job stack. It holds // the instruction pc and the position in the input. type job struct { pc uint32 arg bool pos int } const ( visitedBits = 32 maxBacktrackProg = 500 // len(prog.Inst) <= max
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 14 17:25:39 UTC 2023 - 8.8K bytes - Viewed (0) -
src/runtime/nonwindows_stub.go
func enableWER() {} // winlibcall is not implemented on non-Windows systems, // but it is used in non-OS-specific parts of the runtime. // Define it as an empty struct to avoid wasting stack space.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 26 03:12:13 UTC 2024 - 962 bytes - Viewed (0) -
test/abi/bad_internal_offsets.go
F1 string F2 StructF0S1 } type StructF0S1 struct { _ uint16 } // 0 returns 3 params //go:registerparams //go:noinline func Test0(p0 uint32, p1 StructF0S0, p2 int32) { // consume some stack space, so as to trigger morestack var pad [256]uint64 pad[FailCount]++ if p0 == 0 { return } p1f0c := int16(-3096) if p1.F0 != p1f0c { NoteFailureElem(0, "genChecker0", "parm", 1, 0, pad[0]) return
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 19 23:33:25 UTC 2023 - 1.4K bytes - Viewed (0) -
src/internal/trace/traceviewer/format/format.go
Time float64 `json:"ts"` Dur float64 `json:"dur,omitempty"` PID uint64 `json:"pid"` TID uint64 `json:"tid"` ID uint64 `json:"id,omitempty"` BindPoint string `json:"bp,omitempty"` Stack int `json:"sf,omitempty"` EndStack int `json:"esf,omitempty"` Arg any `json:"args,omitempty"` Cname string `json:"cname,omitempty"` Category string `json:"cat,omitempty"` }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Nov 21 20:45:06 UTC 2023 - 2K bytes - Viewed (0)