Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for asanenabled (0.12 sec)

  1. src/runtime/select.go

    		} else if cas.elem != nil {
    			raceWriteObjectPC(c.elemtype, cas.elem, casePC(casi), chanrecvpc)
    		}
    	}
    	if msanenabled {
    		if casi < nsends {
    			msanread(cas.elem, c.elemtype.Size_)
    		} else if cas.elem != nil {
    			msanwrite(cas.elem, c.elemtype.Size_)
    		}
    	}
    	if asanenabled {
    		if casi < nsends {
    			asanread(cas.elem, c.elemtype.Size_)
    		} else if cas.elem != nil {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Mar 13 21:36:04 UTC 2024
    - 15K bytes
    - Viewed (0)
  2. src/runtime/slice.go

    		}
    	}
    
    	if raceenabled {
    		callerpc := getcallerpc()
    		pc := abi.FuncPCABIInternal(makeslicecopy)
    		racereadrangepc(from, copymem, callerpc, pc)
    	}
    	if msanenabled {
    		msanread(from, copymem)
    	}
    	if asanenabled {
    		asanread(from, copymem)
    	}
    
    	memmove(to, from, copymem)
    
    	return to
    }
    
    // makeslice should be an internal detail,
    // but widely used packages access it using linkname.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 16:25:21 UTC 2024
    - 12.2K bytes
    - Viewed (0)
  3. src/runtime/string.go

    		return ""
    	}
    	if raceenabled {
    		racereadrangepc(unsafe.Pointer(ptr),
    			uintptr(n),
    			getcallerpc(),
    			abi.FuncPCABIInternal(slicebytetostring))
    	}
    	if msanenabled {
    		msanread(unsafe.Pointer(ptr), uintptr(n))
    	}
    	if asanenabled {
    		asanread(unsafe.Pointer(ptr), uintptr(n))
    	}
    	if n == 1 {
    		p := unsafe.Pointer(&staticuint64s[*ptr])
    		if goarch.BigEndian {
    			p = add(p, 7)
    		}
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:17:26 UTC 2024
    - 13.4K bytes
    - Viewed (0)
  4. src/runtime/mbarrier.go

    		raceReadObjectPC(typ, src, getcallerpc(), abi.FuncPCABIInternal(reflect_typedmemmove))
    	}
    	if msanenabled {
    		msanwrite(dst, typ.Size_)
    		msanread(src, typ.Size_)
    	}
    	if asanenabled {
    		asanwrite(dst, typ.Size_)
    		asanread(src, typ.Size_)
    	}
    	typedmemmove(typ, dst, src)
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 15.7K bytes
    - Viewed (0)
  5. src/runtime/iface.go

    func convT(t *_type, v unsafe.Pointer) unsafe.Pointer {
    	if raceenabled {
    		raceReadObjectPC(t, v, getcallerpc(), abi.FuncPCABIInternal(convT))
    	}
    	if msanenabled {
    		msanread(v, t.Size_)
    	}
    	if asanenabled {
    		asanread(v, t.Size_)
    	}
    	x := mallocgc(t.Size_, t, true)
    	typedmemmove(t, x, v)
    	return x
    }
    func convTnoptr(t *_type, v unsafe.Pointer) unsafe.Pointer {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 29 17:58:53 UTC 2024
    - 22.5K bytes
    - Viewed (0)
  6. src/runtime/arena.go

    	}
    
    	// Put the active chunk onto the reuse list.
    	//
    	// Note that active's reference is always the last reference in refs.
    	s = a.active
    	if s != nil {
    		if raceenabled || msanenabled || asanenabled {
    			// Don't reuse arenas with sanitizers enabled. We want to catch
    			// any use-after-free errors aggressively.
    			freeUserArenaChunk(s, a.refs[len(a.refs)-1])
    		} else {
    			lock(&userArenaState.lock)
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:44:56 UTC 2024
    - 37.9K bytes
    - Viewed (0)
  7. src/runtime/mgcsweep.go

    				siter.next()
    			}
    		}
    	}
    	if hadSpecials && s.specials == nil {
    		spanHasNoSpecials(s)
    	}
    
    	if traceAllocFreeEnabled() || debug.clobberfree != 0 || raceenabled || msanenabled || asanenabled {
    		// Find all newly freed objects.
    		mbits := s.markBitsForBase()
    		abits := s.allocBitsForIndex(0)
    		for i := uintptr(0); i < uintptr(s.nelems); i++ {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 08 17:52:18 UTC 2024
    - 32.9K bytes
    - Viewed (0)
  8. src/runtime/stack.go

    		if trace.ok() {
    			trace.GoroutineStackAlloc(uintptr(v), uintptr(n))
    			traceRelease(trace)
    		}
    	}
    	if raceenabled {
    		racemalloc(v, uintptr(n))
    	}
    	if msanenabled {
    		msanmalloc(v, uintptr(n))
    	}
    	if asanenabled {
    		asanunpoison(v, uintptr(n))
    	}
    	if stackDebug >= 1 {
    		print("  allocated ", v, "\n")
    	}
    	return stack{uintptr(v), uintptr(v) + uintptr(n)}
    }
    
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 22 22:31:00 UTC 2024
    - 41.1K bytes
    - Viewed (0)
Back to top