- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 41 for Atack (0.06 sec)
-
src/cmd/link/internal/ld/elf.go
if ctxt.IsMIPS() { shstrtabAddstring(elfRelType + ".MIPS.abiflags") shstrtabAddstring(elfRelType + ".gnu.attributes") } // add a .note.GNU-stack section to mark the stack as non-executable shstrtabAddstring(".note.GNU-stack") if ctxt.IsShared() { shstrtabAddstring(".note.go.abihash") shstrtabAddstring(".note.go.pkg-list") shstrtabAddstring(".note.go.deps") } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Apr 22 13:29:54 UTC 2024 - 63.6K bytes - Viewed (0) -
src/cmd/link/internal/ld/xcoff.go
Odatapsize uint8 // Requested data page size Ostackpsize uint8 // Requested stack page size Oflags uint8 // Flags And TLS Alignment Otsize uint64 // Text Size In Bytes Odsize uint64 // Data Size In Bytes Obsize uint64 // Bss Size In Bytes Oentry uint64 // Entry Point Address Omaxstack uint64 // Max Stack Size Allowed Omaxdata uint64 // Max Data Size Allowed
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Dec 01 19:58:23 UTC 2023 - 51.8K bytes - Viewed (0) -
src/runtime/mheap.go
// whether the span has been zeroed. Note that it may not be. func (h *mheap) alloc(npages uintptr, spanclass spanClass) *mspan { // Don't do any operations that lock the heap on the G stack. // It might trigger stack growth, and the stack growth code needs // to be able to allocate heap. var s *mspan systemstack(func() { // To prevent excessive heap growth, before allocating n pages
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/prove.go
// Called when backing up on a branch. func (ft *factsTable) restore() { if ft.unsatDepth > 0 { ft.unsatDepth-- } else { ft.unsat = false } for { old := ft.stack[len(ft.stack)-1] ft.stack = ft.stack[:len(ft.stack)-1] if old == checkpointFact { break } if old.r == lt|eq|gt { delete(ft.facts, old.p) } else { ft.facts[old.p] = old.r } } for {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:30:21 UTC 2024 - 48.9K bytes - Viewed (0) -
src/cmd/dist/build.go
// Do not include local development, so that people working in the // main branch for day-to-day work on the Go toolchain itself can // still have full paths for stack traces for compiler crashes and the like. env = append(env, "GOFLAGS=-trimpath -ldflags=-w -gcflags=cmd/...=-dwarf=false") } return env }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 18:34:40 UTC 2024 - 54K bytes - Viewed (0) -
src/go/printer/testdata/parser.go
imports []*ast.ImportSpec // list of imports // Label scope // (maintained by open/close LabelScope) labelScope *ast.Scope // label scope for current function targetStack [][]*ast.Ident // stack of unresolved labels } // scannerMode returns the scanner mode bits given the parser's mode bits. func scannerMode(mode uint) uint { var m uint = scanner.InsertSemis if mode&ParseComments != 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Jul 20 20:19:51 UTC 2023 - 50.5K bytes - Viewed (0) -
src/index/suffixarray/sais2.go
// max(maxID, numLMS/2). This level of the recursion needs maxID, // and all deeper levels of the recursion will need no more than numLMS/2, // so this one allocation is guaranteed to suffice for the entire stack // of recursive calls. tmp := oldTmp if len(tmp) < len(saTmp) { tmp = saTmp } if len(tmp) < numLMS { // TestSAIS/forcealloc reaches this code. n := maxID if n < numLMS/2 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 18 23:57:18 UTC 2024 - 52.3K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/sys/unix/syscall_linux.go
// minIovec is the size of the small initial allocation used by // Readv, Writev, etc. // // This small allocation gets stack allocated, which lets the // common use case of len(iovs) <= minIovs avoid more expensive // heap allocations. const minIovec = 8 // appendBytes converts bs to Iovecs and appends them to vecs.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Mar 07 05:26:45 UTC 2024 - 77.5K bytes - Viewed (0) -
src/go/parser/parser.go
p.nestLev++ if p.nestLev > maxNestLev { p.error(p.pos, "exceeded max nesting depth") panic(bailout{}) } return p } // decNestLev is used to track nesting depth during parsing to prevent stack exhaustion. // It is used along with incNestLev in a similar fashion to how un and trace are used. func decNestLev(p *parser) { p.nestLev-- } // Advance to the next token. func (p *parser) next0() {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Dec 08 20:07:50 UTC 2023 - 72.2K bytes - Viewed (0) -
src/cmd/link/internal/ppc64/asm.go
// // There are 3 cases today (as paraphrased from the ELFv2 document): // // 1. R2 holds the TOC pointer on entry. The call stub must save R2 into the ELFv2 TOC stack save slot. // // 2. R2 holds the TOC pointer on entry. The caller has already saved R2 to the TOC stack save slot. // // 3. R2 does not hold the TOC pointer on entry. The caller has no expectations of R2. //
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 19 20:54:08 UTC 2024 - 63.7K bytes - Viewed (0)