- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 120 for uptr (3.05 sec)
-
src/runtime/race_ppc64le.s
GO_ARGS // void __tsan_go_atomic32_load(ThreadState *thr, uptr cpc, uptr pc, u8 *a); MOVD $__tsan_go_atomic32_load(SB), R8 ADD $32, R1, R6 // addr of caller's 1st arg BR racecallatomic<>(SB) RET TEXT sync∕atomic·LoadInt64(SB), NOSPLIT, $0-16 GO_ARGS // void __tsan_go_atomic64_load(ThreadState *thr, uptr cpc, uptr pc, u8 *a); MOVD $__tsan_go_atomic64_load(SB), R8
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 18:37:29 UTC 2024 - 17K bytes - Viewed (0) -
src/runtime/map_faststr.go
if h.B == 0 { // One-bucket table. b := (*bmap)(h.buckets) if key.len < 32 { // short key, doing lots of comparisons is ok for i, kptr := uintptr(0), b.keys(); i < abi.MapBucketCount; i, kptr = i+1, add(kptr, 2*goarch.PtrSize) { k := (*stringStruct)(kptr) if k.len != key.len || isEmpty(b.tophash[i]) { if b.tophash[i] == emptyRest { break } continue }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 15.3K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/generic.rules
(NilCheck ptr:(Addr {_} (SB)) _) => ptr (NilCheck ptr:(Convert (Addr {_} (SB)) _) _) => ptr // for late-expanded calls, recognize memequal applied to a single constant byte // Support is limited by 1, 2, 4, 8 byte sizes (StaticLECall {callAux} sptr (Addr {scon} (SB)) (Const64 [1]) mem) && isSameCall(callAux, "runtime.memequal") && symIsRO(scon)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 16 22:21:05 UTC 2024 - 135.3K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_wasm.go
*ptr = new return old } //go:nosplit //go:noinline func Xchgint32(ptr *int32, new int32) int32 { old := *ptr *ptr = new return old } //go:nosplit //go:noinline func Xchgint64(ptr *int64, new int64) int64 { old := *ptr *ptr = new return old } //go:nosplit //go:noinline func Xchguintptr(ptr *uintptr, new uintptr) uintptr { old := *ptr *ptr = new
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 15 19:57:43 UTC 2024 - 5.4K bytes - Viewed (0) -
src/runtime/atomic_pointer.go
// //go:linkname atomicwb //go:nosplit func atomicwb(ptr *unsafe.Pointer, new unsafe.Pointer) { slot := (*uintptr)(unsafe.Pointer(ptr)) buf := getg().m.p.ptr().wbBuf.get2() buf[0] = *slot buf[1] = uintptr(new) } // atomicstorep performs *ptr = new atomically and invokes a write barrier. // //go:nosplit func atomicstorep(ptr unsafe.Pointer, new unsafe.Pointer) { if writeBarrier.enabled {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 4K bytes - Viewed (0) -
src/reflect/value.go
switch v.kind() { case Bool: *(*bool)(v.ptr) = false case Int: *(*int)(v.ptr) = 0 case Int8: *(*int8)(v.ptr) = 0 case Int16: *(*int16)(v.ptr) = 0 case Int32: *(*int32)(v.ptr) = 0 case Int64: *(*int64)(v.ptr) = 0 case Uint: *(*uint)(v.ptr) = 0 case Uint8: *(*uint8)(v.ptr) = 0 case Uint16: *(*uint16)(v.ptr) = 0 case Uint32: *(*uint32)(v.ptr) = 0
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 21:17:41 UTC 2024 - 119.9K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_andor_generic.go
import _ "unsafe" // For linkname //go:nosplit func And32(ptr *uint32, val uint32) uint32 { for { old := *ptr if Cas(ptr, old, old&val) { return old } } } //go:nosplit func Or32(ptr *uint32, val uint32) uint32 { for { old := *ptr if Cas(ptr, old, old|val) { return old } } } //go:nosplit func And64(ptr *uint64, val uint64) uint64 { for { old := *ptr
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 20:08:37 UTC 2024 - 1.2K bytes - Viewed (0) -
cmd/kubeadm/app/util/config/upgradeconfiguration_test.go
CertificateRenewal: ptr.To(true), EtcdUpgrade: ptr.To(true), ImagePullPolicy: v1.PullIfNotPresent, ImagePullSerial: ptr.To(true), }, Node: kubeadmapi.UpgradeNodeConfiguration{ CertificateRenewal: ptr.To(true), EtcdUpgrade: ptr.To(true), ImagePullPolicy: v1.PullIfNotPresent, ImagePullSerial: ptr.To(true), }, }, }, {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Thu May 16 08:34:39 UTC 2024 - 11.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64.rules
(MOVHstore ptr (MOVDconst [0]) mem)) (Zero [5] ptr mem) => (MOVBstore [4] ptr (MOVDconst [0]) (MOVWstore ptr (MOVDconst [0]) mem)) (Zero [6] ptr mem) => (MOVHstore [4] ptr (MOVDconst [0]) (MOVWstore ptr (MOVDconst [0]) mem)) (Zero [7] ptr mem) => (MOVWstore [3] ptr (MOVDconst [0]) (MOVWstore ptr (MOVDconst [0]) mem)) (Zero [8] ptr mem) => (MOVDstore ptr (MOVDconst [0]) mem) (Zero [9] ptr mem) =>
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 113.1K bytes - Viewed (0) -
src/internal/runtime/atomic/atomic_mipsx.go
func Loadp(ptr unsafe.Pointer) unsafe.Pointer //go:noescape func LoadAcq(ptr *uint32) uint32 //go:noescape func LoadAcquintptr(ptr *uintptr) uintptr //go:noescape func And8(ptr *uint8, val uint8) //go:noescape func Or8(ptr *uint8, val uint8) //go:noescape func And(ptr *uint32, val uint32) //go:noescape func Or(ptr *uint32, val uint32) //go:noescape
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 17 20:08:37 UTC 2024 - 3.2K bytes - Viewed (0)