- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 735 for cbits (0.06 sec)
-
src/runtime/cgocheck.go
// bytes, and throws if it finds an unpinned Go pointer. The gcbits mark each // pointer value. The src pointer is off bytes into the gcbits. // //go:nosplit //go:nowritebarrier func cgoCheckBits(src unsafe.Pointer, gcbits *byte, off, size uintptr) { skipMask := off / goarch.PtrSize / 8 skipBytes := skipMask * goarch.PtrSize * 8 ptrmask := addb(gcbits, skipMask) src = add(src, skipBytes) off -= skipBytes
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 09 04:07:57 UTC 2024 - 7.6K bytes - Viewed (0) -
src/runtime/mcheckmark.go
} // setCheckmark throws if marking object is a checkmarks violation, // and otherwise sets obj's checkmark. It returns true if obj was // already checkmarked. func setCheckmark(obj, base, off uintptr, mbits markBits) bool { if !mbits.isMarked() { printlock() print("runtime: checkmarks found unexpected unmarked object obj=", hex(obj), "\n") print("runtime: found obj at *(", hex(base), "+", hex(off), ")\n")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 2.8K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/rewrite.go
me = 32 - bits.TrailingZeros32(uint32(mask)) mbn = bits.LeadingZeros32(^uint32(mask)) men = 32 - bits.TrailingZeros32(^uint32(mask)) } else { mb = bits.LeadingZeros64(uint64(mask)) me = 64 - bits.TrailingZeros64(uint64(mask)) mbn = bits.LeadingZeros64(^uint64(mask)) men = 64 - bits.TrailingZeros64(^uint64(mask)) } // Check for a wrapping mask (e.g bits at 0 and 63)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 64.2K bytes - Viewed (0) -
src/hash/crc32/gen_const_ppc64le.go
} } func genCrc32ConstTable(w *bytes.Buffer, poly uint32, polyid string) { ref_poly := reflect_bits(uint64(poly), 32) fmt.Fprintf(w, "\n\t/* Reduce %d kbits to 1024 bits */\n", blocking*8) j := 0 for i := (blocking * 8) - 1024; i > 0; i -= 1024 { a := reflect_bits(get_remainder(ref_poly, 32, uint(i)), 32) << 1 b := reflect_bits(get_remainder(ref_poly, 32, uint(i+64)), 32) << 1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 19 20:44:20 UTC 2024 - 4.5K bytes - Viewed (0) -
src/runtime/pinner.go
atomic.And8(v.bytep, ^mask) } } // pinnerBits is the same type as gcBits but has different methods. type pinnerBits gcBits // ofObject returns the pinState of the n'th object. // nosplit, because it's called by isPinned, which is nosplit // //go:nosplit func (p *pinnerBits) ofObject(n uintptr) pinState { bytep, mask := (*gcBits)(p).bitp(n * 2) byteVal := atomic.Load8(bytep) return pinState{bytep, byteVal, mask}
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 04 14:29:45 UTC 2024 - 11K bytes - Viewed (0) -
src/cmd/internal/obj/riscv/obj.go
// Bit order - [8|4:3|7:6|2:1|5] bits := extractBitAndShift(imm, 8, 7) bits |= extractBitAndShift(imm, 4, 6) bits |= extractBitAndShift(imm, 3, 5) bits |= extractBitAndShift(imm, 7, 4) bits |= extractBitAndShift(imm, 6, 3) bits |= extractBitAndShift(imm, 2, 2) bits |= extractBitAndShift(imm, 1, 1) bits |= extractBitAndShift(imm, 5, 0) return (bits>>5)<<10 | (bits&0x1f)<<2 }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Sun Apr 07 03:32:27 UTC 2024 - 77K bytes - Viewed (0) -
src/runtime/mheap.go
} // newAllocBits returns a pointer to 8 byte aligned bytes // to be used for this span's alloc bits. // newAllocBits is used to provide newly initialized spans // allocation bits. For spans not being initialized the // mark bits are repurposed as allocation bits when // the span is swept. func newAllocBits(nelems uintptr) *gcBits { return newMarkBits(nelems) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 22 22:31:00 UTC 2024 - 78K bytes - Viewed (0) -
src/runtime/mwbbuf.go
obj, span, objIndex := findObject(ptr, 0, 0) if obj == 0 { continue } // TODO: Consider making two passes where the first // just prefetches the mark bits. mbits := span.markBitsForIndex(objIndex) if mbits.isMarked() { continue } mbits.setMarked() // Mark span. arena, pageIdx, pageMask := pageIndexOf(span.base()) if arena.pageMarks[pageIdx]&pageMask == 0 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.1K bytes - Viewed (0) -
src/math/big/nat.go
// That is, we can throw away all but the bottom logM-1 bits of y. // Instead of allocating a new y, we start reading y at the right word // and truncate it appropriately at the start of the loop. i := len(y) - 1 mtop := int((logM - 2) / _W) // -2 because the top word of N bits is the (N-1)/W'th word. mmask := ^Word(0) if mbits := (logM - 1) & (_W - 1); mbits != 0 { mmask = (1 << mbits) - 1 } if i > mtop { i = mtop
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 13 21:31:58 UTC 2024 - 31.7K bytes - Viewed (0) -
src/runtime/mgcmark.go
// base and extent. b := b0 n := n0 for i := uintptr(0); i < n; { // Find bits for the next word. bits := uint32(*addb(ptrmask, i/(goarch.PtrSize*8))) if bits == 0 { i += goarch.PtrSize * 8 continue } for j := 0; j < 8 && i < n; j++ { if bits&1 != 0 { // Same work as in scanobject; see comments there. p := *(*uintptr)(unsafe.Pointer(b + i))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 21:25:11 UTC 2024 - 52.5K bytes - Viewed (0)