- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 17 for msanenabled (0.14 sec)
-
src/runtime/msan.go
func MSanRead(addr unsafe.Pointer, len int) { msanread(addr, uintptr(len)) } func MSanWrite(addr unsafe.Pointer, len int) { msanwrite(addr, uintptr(len)) } // Private interface for the runtime. const msanenabled = true // If we are running on the system stack, the C program may have // marked part of that stack as uninitialized. We don't instrument // the runtime, but operations like a slice copy can call msanread
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Feb 20 20:50:21 UTC 2024 - 1.6K bytes - Viewed (0) -
src/runtime/slice.go
} } if raceenabled { callerpc := getcallerpc() pc := abi.FuncPCABIInternal(makeslicecopy) racereadrangepc(from, copymem, callerpc, pc) } if msanenabled { msanread(from, copymem) } if asanenabled { asanread(from, copymem) } memmove(to, from, copymem) return to } // makeslice should be an internal detail, // but widely used packages access it using linkname.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 16:25:21 UTC 2024 - 12.2K bytes - Viewed (0) -
src/runtime/select.go
} else if cas.elem != nil { raceWriteObjectPC(c.elemtype, cas.elem, casePC(casi), chanrecvpc) } } if msanenabled { if casi < nsends { msanread(cas.elem, c.elemtype.Size_) } else if cas.elem != nil { msanwrite(cas.elem, c.elemtype.Size_) } } if asanenabled { if casi < nsends { asanread(cas.elem, c.elemtype.Size_) } else if cas.elem != nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Mar 13 21:36:04 UTC 2024 - 15K bytes - Viewed (0) -
src/runtime/string.go
return "" } if raceenabled { racereadrangepc(unsafe.Pointer(ptr), uintptr(n), getcallerpc(), abi.FuncPCABIInternal(slicebytetostring)) } if msanenabled { msanread(unsafe.Pointer(ptr), uintptr(n)) } if asanenabled { asanread(unsafe.Pointer(ptr), uintptr(n)) } if n == 1 { p := unsafe.Pointer(&staticuint64s[*ptr]) if goarch.BigEndian { p = add(p, 7) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 01:17:26 UTC 2024 - 13.4K bytes - Viewed (0) -
src/runtime/mbarrier.go
raceReadObjectPC(typ, src, getcallerpc(), abi.FuncPCABIInternal(reflect_typedmemmove)) } if msanenabled { msanwrite(dst, typ.Size_) msanread(src, typ.Size_) } if asanenabled { asanwrite(dst, typ.Size_) asanread(src, typ.Size_) } typedmemmove(typ, dst, src) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 15.7K bytes - Viewed (0) -
src/runtime/iface.go
func convT(t *_type, v unsafe.Pointer) unsafe.Pointer { if raceenabled { raceReadObjectPC(t, v, getcallerpc(), abi.FuncPCABIInternal(convT)) } if msanenabled { msanread(v, t.Size_) } if asanenabled { asanread(v, t.Size_) } x := mallocgc(t.Size_, t, true) typedmemmove(t, x, v) return x } func convTnoptr(t *_type, v unsafe.Pointer) unsafe.Pointer {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 22.5K bytes - Viewed (0) -
src/runtime/arena.go
} // Put the active chunk onto the reuse list. // // Note that active's reference is always the last reference in refs. s = a.active if s != nil { if raceenabled || msanenabled || asanenabled { // Don't reuse arenas with sanitizers enabled. We want to catch // any use-after-free errors aggressively. freeUserArenaChunk(s, a.refs[len(a.refs)-1]) } else { lock(&userArenaState.lock)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:44:56 UTC 2024 - 37.9K bytes - Viewed (0) -
src/runtime/map.go
callerpc := getcallerpc() pc := abi.FuncPCABIInternal(mapaccess1) racereadpc(unsafe.Pointer(h), callerpc, pc) raceReadObjectPC(t.Key, key, callerpc, pc) } if msanenabled && h != nil { msanread(key, t.Key.Size_) } if asanenabled && h != nil { asanread(key, t.Key.Size_) } if h == nil || h.count == 0 { if err := mapKeyError(t, key); err != nil { panic(err) // see issue 23734 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 29 17:58:53 UTC 2024 - 57.6K bytes - Viewed (0) -
src/runtime/mgcsweep.go
siter.next() } } } if hadSpecials && s.specials == nil { spanHasNoSpecials(s) } if traceAllocFreeEnabled() || debug.clobberfree != 0 || raceenabled || msanenabled || asanenabled { // Find all newly freed objects. mbits := s.markBitsForBase() abits := s.allocBitsForIndex(0) for i := uintptr(0); i < uintptr(s.nelems); i++ {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 17:52:18 UTC 2024 - 32.9K bytes - Viewed (0) -
src/runtime/stack.go
if trace.ok() { trace.GoroutineStackAlloc(uintptr(v), uintptr(n)) traceRelease(trace) } } if raceenabled { racemalloc(v, uintptr(n)) } if msanenabled { msanmalloc(v, uintptr(n)) } if asanenabled { asanunpoison(v, uintptr(n)) } if stackDebug >= 1 { print(" allocated ", v, "\n") } return stack{uintptr(v), uintptr(v) + uintptr(n)} }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 41.1K bytes - Viewed (0)