- Sort Score
- Result 10 results
- Languages All
Results 51 - 60 of 68 for __bits (0.33 sec)
-
tensorflow/compiler/mlir/quantization/tensorflow/python/integration_test/quantize_model_test_base.py
is_qat_model: bool = False, ) -> module.Module: class EinsumModel(module.Module): """Einsum class.""" def __init__(self): self._bias = None if bias_shape is not None: self._bias = array_ops.constant( np.random.uniform(size=bias_shape), dtype=dtypes.float32 ) self._kernel = np.random.uniform(size=y_shape).astype('f4')
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 21 08:51:46 UTC 2024 - 51.2K bytes - Viewed (0) -
src/cmd/compile/internal/types2/sizes.go
// types are naturally aligned with a maximum alignment MaxAlign. // // *StdSizes implements Sizes. type StdSizes struct { WordSize int64 // word size in bytes - must be >= 4 (32bits) MaxAlign int64 // maximum alignment in bytes - must be >= 1 } func (s *StdSizes) Alignof(T Type) (result int64) { defer func() { assert(result >= 1) }()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 8.8K bytes - Viewed (0) -
src/runtime/mgcmark.go
// base and extent. b := b0 n := n0 for i := uintptr(0); i < n; { // Find bits for the next word. bits := uint32(*addb(ptrmask, i/(goarch.PtrSize*8))) if bits == 0 { i += goarch.PtrSize * 8 continue } for j := 0; j < 8 && i < n; j++ { if bits&1 != 0 { // Same work as in scanobject; see comments there. p := *(*uintptr)(unsafe.Pointer(b + i))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Apr 18 21:25:11 UTC 2024 - 52.5K bytes - Viewed (0) -
src/cmd/link/internal/ld/dwarf.go
} // addDwarfAddrField adds a DWARF field in DWARF 64bits or 32bits. func (d *dwctxt) addDwarfAddrField(sb *loader.SymbolBuilder, v uint64) { if isDwarf64(d.linkctxt) { sb.AddUint(d.arch, v) } else { sb.AddUint32(d.arch, uint32(v)) } } // addDwarfAddrRef adds a DWARF pointer in DWARF 64bits or 32bits. func (d *dwctxt) addDwarfAddrRef(sb *loader.SymbolBuilder, t loader.Sym) {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 02 16:25:18 UTC 2024 - 72.4K bytes - Viewed (0) -
src/go/types/sizes.go
// types are naturally aligned with a maximum alignment MaxAlign. // // *StdSizes implements Sizes. type StdSizes struct { WordSize int64 // word size in bytes - must be >= 4 (32bits) MaxAlign int64 // maximum alignment in bytes - must be >= 1 } func (s *StdSizes) Alignof(T Type) (result int64) { defer func() { assert(result >= 1) }()
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Apr 03 18:48:38 UTC 2024 - 8.9K bytes - Viewed (0) -
src/cmd/link/internal/ld/symtab.go
setCarrierSym(t, s.Sym()) return s.Sym() } var ( symgostring = groupSym("go:string.*", sym.SGOSTRING) symgofunc = groupSym("go:func.*", sym.SGOFUNC) symgcbits = groupSym("runtime.gcbits.*", sym.SGCBITS) ) symgofuncrel := symgofunc if ctxt.UseRelro() { symgofuncrel = groupSym("go:funcrel.*", sym.SGOFUNCRELRO) } symt := ldr.CreateSymForUpdate("runtime.symtab", 0)
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Sep 13 16:29:40 UTC 2023 - 29.2K bytes - Viewed (0) -
src/runtime/gc_test.go
// n must always be a multiple of 8, since gcBits is // always rounded up 8 bytes. for _, n := range []int{8, 16, 32, 64, 128} { b.Run(fmt.Sprintf("bits=%d", n*8), func(b *testing.B) { // Initialize a new byte slice with pseduo-random data. bits := make([]byte, n) rand.Read(bits) b.ResetTimer() for i := 0; i < b.N; i++ { runtime.MSpanCountAlloc(s, bits) } }) } }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed Jun 05 22:33:52 UTC 2024 - 17.6K bytes - Viewed (0) -
src/cmd/asm/internal/asm/testdata/amd64enc_extra.s
VADDPD 8128(DX), Z28, Z29 // 62619d40586a7f VADDPD 8128(DX)(AX*2), Z0, Z29 // 6261fd48586c427f VADDPD 8128(DX)(AX*2), Z29, Z1 // 62f19540584c427f // EVEX: compressed displacement that does not fit into 8bits. VADDPD 2048(DX), X29, X0 // 62f19500588200080000 VADDPD 2048(DX), X1, X29 // 6261f50858aa00080000 VADDPD 2048(DX), X29, X28 // 6261950058a200080000
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 11 18:32:50 UTC 2023 - 57.6K bytes - Viewed (0) -
src/cmd/compile/internal/ssa/_gen/ARM64Ops.go
// insert low width bits of arg0 into the result starting at bit lsb, bits to the left of the inserted bit field are set to the high/sign bit of the inserted bit field, bits to the right are zeroed {name: "SBFIZ", argLength: 1, reg: gp11, asm: "SBFIZ", aux: "ARM64BitField"}, // extract width bits of arg0 starting at bit lsb and insert at low end of result, remaining high bits are set to the high/sign bit of the extracted bitfield
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu May 23 15:49:20 UTC 2024 - 58.8K bytes - Viewed (0) -
src/cmd/link/internal/loader/loader.go
return 0 } // TODO: would it make sense to return an arch-specific // alignment depending on section type? E.g. STEXT => 32, // SDATA => 1, etc? abits := l.align[i] if abits == 0 { return 0 } return int32(1 << (abits - 1)) } // SetSymAlign sets the alignment for a symbol. func (l *Loader) SetSymAlign(i Sym, align int32) { // Reject nonsense alignments.
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 24 20:26:10 UTC 2024 - 81.5K bytes - Viewed (0)