- Sort Score
- Result 10 results
- Languages All
Results 21 - 30 of 200 for clobber (0.12 sec)
-
src/runtime/asm_loong64.s
// to the buffer space in R29. // It clobbers R30 (the linker temp register). // The act of CALLing gcWriteBarrier will clobber R1 (LR). // It does not clobber any other general-purpose registers, // but may clobber others (e.g., floating point registers). TEXT gcWriteBarrier<>(SB),NOSPLIT,$216 // Save the registers clobbered by the fast path. MOVV R19, 208(R3) MOVV R13, 216(R3) retry:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 15:04:25 UTC 2024 - 26.5K bytes - Viewed (0) -
src/runtime/asm_mipsx.s
// to the buffer space in R25. // It clobbers R23 (the linker temp register). // The act of CALLing gcWriteBarrier will clobber R31 (LR). // It does not clobber any other general-purpose registers, // but may clobber others (e.g., floating point registers). TEXT gcWriteBarrier<>(SB),NOSPLIT,$104 // Save the registers clobbered by the fast path. MOVW R1, 100(R29) MOVW R2, 104(R29) retry:
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 06 11:46:29 UTC 2024 - 26.3K bytes - Viewed (0) -
src/runtime/asm_arm64.s
// to the buffer space in R25. // It clobbers condition codes. // It does not clobber any general-purpose registers except R27, // but may clobber others (e.g., floating point registers) // The act of CALLing gcWriteBarrier will clobber R30 (LR). TEXT gcWriteBarrier<>(SB),NOSPLIT,$200 // Save the registers clobbered by the fast path. STP (R0, R1), 184(RSP) retry: MOVD g_m(g), R0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 20:38:24 UTC 2024 - 43.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64.rules
// madd/msub (ADD a l:(MUL x y)) && l.Uses==1 && clobber(l) => (MADD a x y) (SUB a l:(MUL x y)) && l.Uses==1 && clobber(l) => (MSUB a x y) (ADD a l:(MNEG x y)) && l.Uses==1 && clobber(l) => (MSUB a x y) (SUB a l:(MNEG x y)) && l.Uses==1 && clobber(l) => (MADD a x y) (ADD a l:(MULW x y)) && v.Type.Size() <= 4 && l.Uses==1 && clobber(l) => (MADDW a x y)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 113.1K bytes - Viewed (0) -
src/runtime/asm_amd64.s
// gcWriteBarrier does NOT follow the Go ABI. It accepts the // number of bytes of buffer needed in R11, and returns a pointer // to the buffer space in R11. // It clobbers FLAGS. It does not clobber any general-purpose registers, // but may clobber others (e.g., SSE registers). // Typical use would be, when doing *(CX+88) = AX // CMPL $0, runtime.writeBarrier(SB) // JEQ dowrite
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sat May 11 20:38:24 UTC 2024 - 60.4K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/PPC64.rules
(MFVSRD (FMOVDconst [c])) => (MOVDconst [int64(math.Float64bits(c))]) (MTVSRD x:(MOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (FMOVDload [off] {sym} ptr mem) (MFVSRD x:(FMOVDload [off] {sym} ptr mem)) && x.Uses == 1 && clobber(x) => @x.Block (MOVDload [off] {sym} ptr mem) // Rules for MOV* or FMOV* ops determine when indexed (MOV*loadidx or MOV*storeidx)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 53.2K bytes - Viewed (0) -
src/runtime/asm_ppc64x.s
// number of bytes of buffer needed in R29, and returns a pointer // to the buffer space in R29. // It clobbers condition codes. // It does not clobber R0 through R17 (except special registers), // but may clobber any other register, *including* R31. TEXT gcWriteBarrier<>(SB),NOSPLIT,$120 // The standard prologue clobbers R31. // We use R18, R19, and R31 as scratch registers. retry: MOVD g_m(g), R18 MOVD m_p(R18), R18
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 18:17:17 UTC 2024 - 45.4K bytes - Viewed (0) -
tensorflow/c/experimental/saved_model/core/tf_saved_model_api.cc
SavedModelV2Bundle bundle; TF_RETURN_IF_ERROR(SavedModelV2Bundle::Load(directory, &bundle)); // TODO(bmzhao): Mangle loaded function names so that different // models loaded in the same runtime Context don't clobber eachother. // This occurs in python here: // https://github.com/tensorflow/tensorflow/blob/285b5fa15405c5e2c084080f52a1818be8648079/tensorflow/python/saved_model/function_deserialization.py#L438-L454
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Feb 27 09:34:33 UTC 2024 - 14.1K bytes - Viewed (0) -
src/runtime/mwbbuf.go
// file deals only with the buffer. // // The write barrier has a fast path and a slow path. The fast path // simply enqueues to a per-P write barrier buffer. It's written in // assembly and doesn't clobber any general purpose registers, so it // doesn't have the usual overheads of a Go call. // // When the buffer fills up, the write barrier invokes the slow path // (wbBufFlush) to flush the buffer to the GC work queues. In this
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/go/types/operand.go
} // x's type V is a type parameter and T is not a named type, // and values x' of each specific type in V's type set are // assignable to T. if Vp != nil && !hasName(T) { x := *x // don't clobber outer x ok := false code := IncompatibleAssign Vp.is(func(V *term) bool { if V == nil { return false // no specific types } x.typ = V.typ ok, code = x.assignableTo(check, T, cause)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 21:17:10 UTC 2024 - 11.1K bytes - Viewed (0)