- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 275 for clobber (0.12 sec)
-
src/cmd/vendor/golang.org/x/tools/internal/facts/facts.go
s.pkg, obj, fact) } key := key{pkg: obj.Pkg(), obj: obj, t: reflect.TypeOf(fact)} s.mu.Lock() s.m[key] = fact // clobber any existing entry s.mu.Unlock() } func (s *Set) AllObjectFacts(filter map[reflect.Type]bool) []analysis.ObjectFact { var facts []analysis.ObjectFact s.mu.Lock() for k, v := range s.m {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 02:38:00 UTC 2024 - 12.1K bytes - Viewed (0) -
src/runtime/sys_windows_arm64.s
#define TEB_TlsSlots 0x1480 #define TEB_ArbitraryPtr 0x28 // Note: R0-R7 are args, R8 is indirect return value address, // R9-R15 are caller-save, R19-R29 are callee-save. // // load_g and save_g (in tls_arm64.s) clobber R27 (REGTMP) and R0. TEXT runtime·asmstdcall_trampoline<ABIInternal>(SB),NOSPLIT,$0 B runtime·asmstdcall(SB) // void runtime·asmstdcall(void *c); TEXT runtime·asmstdcall(SB),NOSPLIT,$16
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Sep 21 15:56:43 UTC 2023 - 6.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/genericOps.go
{name: "PubBarrier", argLength: 1, hasSideEffects: true}, // Do data barrier. arg0=memory. // Clobber experiment op {name: "Clobber", argLength: 0, typ: "Void", aux: "SymOff", symEffect: "None"}, // write an invalid pointer value to the given pointer slot of a stack variable {name: "ClobberReg", argLength: 0, typ: "Void"}, // clobber a register // Prefetch instruction
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 42.6K bytes - Viewed (0) -
src/runtime/asm_loong64.s
// to the buffer space in R29. // It clobbers R30 (the linker temp register). // The act of CALLing gcWriteBarrier will clobber R1 (LR). // It does not clobber any other general-purpose registers, // but may clobber others (e.g., floating point registers). TEXT gcWriteBarrier<>(SB),NOSPLIT,$216 // Save the registers clobbered by the fast path. MOVV R19, 208(R3) MOVV R13, 216(R3) retry:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 15:04:25 UTC 2024 - 26.5K bytes - Viewed (0) -
src/runtime/asm_mipsx.s
// to the buffer space in R25. // It clobbers R23 (the linker temp register). // The act of CALLing gcWriteBarrier will clobber R31 (LR). // It does not clobber any other general-purpose registers, // but may clobber others (e.g., floating point registers). TEXT gcWriteBarrier<>(SB),NOSPLIT,$104 // Save the registers clobbered by the fast path. MOVW R1, 100(R29) MOVW R2, 104(R29) retry:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 06 11:46:29 UTC 2024 - 26.3K bytes - Viewed (0) -
src/runtime/asm_arm64.s
// to the buffer space in R25. // It clobbers condition codes. // It does not clobber any general-purpose registers except R27, // but may clobber others (e.g., floating point registers) // The act of CALLing gcWriteBarrier will clobber R30 (LR). TEXT gcWriteBarrier<>(SB),NOSPLIT,$200 // Save the registers clobbered by the fast path. STP (R0, R1), 184(RSP) retry: MOVD g_m(g), R0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 20:38:24 UTC 2024 - 43.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64.rules
// madd/msub (ADD a l:(MUL x y)) && l.Uses==1 && clobber(l) => (MADD a x y) (SUB a l:(MUL x y)) && l.Uses==1 && clobber(l) => (MSUB a x y) (ADD a l:(MNEG x y)) && l.Uses==1 && clobber(l) => (MSUB a x y) (SUB a l:(MNEG x y)) && l.Uses==1 && clobber(l) => (MADD a x y) (ADD a l:(MULW x y)) && v.Type.Size() <= 4 && l.Uses==1 && clobber(l) => (MADDW a x y)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 113.1K bytes - Viewed (0) -
src/runtime/asm_amd64.s
// gcWriteBarrier does NOT follow the Go ABI. It accepts the // number of bytes of buffer needed in R11, and returns a pointer // to the buffer space in R11. // It clobbers FLAGS. It does not clobber any general-purpose registers, // but may clobber others (e.g., SSE registers). // Typical use would be, when doing *(CX+88) = AX // CMPL $0, runtime.writeBarrier(SB) // JEQ dowrite
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 20:38:24 UTC 2024 - 60.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/PPC64.rules
(MFVSRD (FMOVDconst [c])) => (MOVDconst [int64(math.Float64bits(c))]) (MTVSRD x:(MOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (FMOVDload [off] {sym} ptr mem) (MFVSRD x:(FMOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVDload [off] {sym} ptr mem) // Rules for MOV* or FMOV* ops determine when indexed (MOV*loadidx or MOV*storeidx)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0) -
src/runtime/asm_ppc64x.s
// number of bytes of buffer needed in R29, and returns a pointer // to the buffer space in R29. // It clobbers condition codes. // It does not clobber R0 through R17 (except special registers), // but may clobber any other register, *including* R31. TEXT gcWriteBarrier<>(SB),NOSPLIT,$120 // The standard prologue clobbers R31. // We use R18, R19, and R31 as scratch registers. retry: MOVD g_m(g), R18 MOVD m_p(R18), R18
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 18:17:17 UTC 2024 - 45.4K bytes - Viewed (0)