- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 32 for clobber (0.14 sec)
-
src/cmd/compile/internal/ssa/_gen/AMD64.rules
(MOVBQZX x:(MOVBload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVBload <v.Type> [off] {sym} ptr mem) (MOVBQZX x:(MOVWload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVBload <v.Type> [off] {sym} ptr mem) (MOVBQZX x:(MOVLload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVBload <v.Type> [off] {sym} ptr mem)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 12 19:38:41 UTC 2024 - 93.9K bytes - Viewed (0) -
src/runtime/asm_amd64.s
// gcWriteBarrier does NOT follow the Go ABI. It accepts the // number of bytes of buffer needed in R11, and returns a pointer // to the buffer space in R11. // It clobbers FLAGS. It does not clobber any general-purpose registers, // but may clobber others (e.g., SSE registers). // Typical use would be, when doing *(CX+88) = AX // CMPL $0, runtime.writeBarrier(SB) // JEQ dowrite
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 20:38:24 UTC 2024 - 60.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/PPC64.rules
(MFVSRD (FMOVDconst [c])) => (MOVDconst [int64(math.Float64bits(c))]) (MTVSRD x:(MOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (FMOVDload [off] {sym} ptr mem) (MFVSRD x:(FMOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVDload [off] {sym} ptr mem) // Rules for MOV* or FMOV* ops determine when indexed (MOV*loadidx or MOV*storeidx)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite.go
return nil // too far away } // clobber invalidates values. Returns true. // clobber is used by rewrite rules to: // // A) make sure the values are really dead and never used again. // B) decrement use counts of the values' args. func clobber(vv ...*Value) bool { for _, v := range vv { v.reset(OpInvalid) // Note: leave v.Block intact. The Block field is used after clobber. } return true }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64Ops.go
// It saves all GP registers if necessary, // but clobbers R30 (LR) because it's a call. // R16 and R17 may be clobbered by linker trampoline. // Returns a pointer to a write barrier buffer in R25. {name: "LoweredWB", argLength: 1, reg: regInfo{clobbers: (callerSave &^ gpg) | buildReg("R16 R17 R30"), outputs: []regMask{buildReg("R25")}}, clobberFlags: true, aux: "Int64"},
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 58.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/debug.go
} // Handle any register clobbering. Call operations, for example, // clobber all registers even though they don't explicitly write to // them. clobbers := uint64(opcodeTable[v.Op].reg.clobbers) for { if clobbers == 0 { break } reg := uint8(bits.TrailingZeros64(clobbers)) clobbers &^= 1 << reg for _, slot := range locs.registers[reg] { if state.loggingLevel > 1 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Jun 10 19:44:43 UTC 2024 - 58.4K bytes - Viewed (0) -
src/crypto/internal/nistec/p256_asm_ppc64le.s
LXVD2X (R16)(CPOOL), P1 LXVD2X (R0)(CPOOL), P0 CALL p256MulInternal<>(SB) MOVD n+16(FP), N ADD $-1, N CMP $0, N BEQ done MOVD N, n+16(FP) // Save counter to avoid clobber VOR T0, T0, X0 VOR T1, T1, X1 BR sqrLoop done: MOVD $p256mul<>+0x00(SB), CPOOL XXPERMDI T0, T0, $2, T0 XXPERMDI T1, T1, $2, T1 STXVD2X T0, (R0)(res_ptr)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 56.5K bytes - Viewed (0) -
src/cmd/link/internal/ppc64/asm.go
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Mar 19 20:54:08 UTC 2024 - 63.7K bytes - Viewed (0) -
src/runtime/mbitmap.go
m.mask = 1 } else { m.mask = m.mask << 1 } m.index++ } // clobberdeadPtr is a special value that is used by the compiler to // clobber dead stack slots, when -clobberdead flag is set. const clobberdeadPtr = uintptr(0xdeaddead | 0xdeaddead<<((^uintptr(0)>>63)*32)) // badPointer throws bad pointer in heap panic.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 00:18:55 UTC 2024 - 60K bytes - Viewed (0) -
guava/src/com/google/common/collect/MapMakerInternalMap.java
// Mimic // "if (!map.containsKey(key)) ... // else return map.get(key); return entryValue; } else { // clobber existing entry, count remains unchanged ++modCount; setValue(e, value); return entryValue; } } }
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Sat May 18 03:24:34 UTC 2024 - 90.8K bytes - Viewed (0)