- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 305 for nbits (0.04 sec)
-
src/compress/lzw/reader_test.go
{11, 2048 - 1024}, {12, 4096 - 2048}, } nCodes, nBits := 0, 0 for _, e := range iterations { nCodes += e.n nBits += e.n * e.width } if nCodes != 3839 { t.Fatalf("nCodes: got %v, want %v", nCodes, 3839) } if nBits != 43255 { t.Fatalf("nBits: got %v, want %v", nBits, 43255) } // Construct our input of 43255 zero bits (which gets d.hi and d.width up
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 18 16:57:58 UTC 2024 - 7.6K bytes - Viewed (0) -
src/compress/lzw/writer.go
} // writeLSB writes the code c for "Least Significant Bits first" data. func (w *Writer) writeLSB(c uint32) error { w.bits |= c << w.nBits w.nBits += w.width for w.nBits >= 8 { if err := w.w.WriteByte(uint8(w.bits)); err != nil { return err } w.bits >>= 8 w.nBits -= 8 } return nil } // writeMSB writes the code c for "Most Significant Bits first" data. func (w *Writer) writeMSB(c uint32) error {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Apr 26 13:32:40 UTC 2024 - 7.9K bytes - Viewed (0) -
src/compress/lzw/reader.go
} // readLSB returns the next code for "Least Significant Bits first" data. func (r *Reader) readLSB() (uint16, error) { for r.nBits < r.width { x, err := r.r.ReadByte() if err != nil { return 0, err } r.bits |= uint32(x) << r.nBits r.nBits += 8 } code := uint16(r.bits & (1<<r.width - 1)) r.bits >>= r.width r.nBits -= r.width return code, nil }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Oct 12 14:39:39 UTC 2023 - 8K bytes - Viewed (0) -
src/internal/zstd/bits.go
cnt uint32 // number of valid bits in the bits field } // makeBitReader makes a bit reader starting at off. func (r *Reader) makeBitReader(data block, off int) bitReader { return bitReader{ r: r, data: data, off: uint32(off), } } // moreBits is called to read more bits. // This ensures that at least 16 bits are available. func (br *bitReader) moreBits() error { for br.cnt < 16 {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Tue Apr 18 20:34:13 UTC 2023 - 3.3K bytes - Viewed (0) -
test/codegen/bits.go
func bitSetTest(x int) bool { // amd64:"ANDL\t[$]9, AX" // amd64:"CMPQ\tAX, [$]9" return x&9 == 9 } // mask contiguous one bits func cont1Mask64U(x uint64) uint64 { // s390x:"RISBGZ\t[$]16, [$]47, [$]0," return x & 0x0000ffffffff0000 } // mask contiguous zero bits func cont0Mask64U(x uint64) uint64 { // s390x:"RISBGZ\t[$]48, [$]15, [$]0," return x & 0xffff00000000ffff }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Jun 07 19:02:52 UTC 2024 - 7.8K bytes - Viewed (0) -
src/vendor/golang.org/x/net/http2/hpack/huffman.go
// cur is the bit buffer that has not been fed into n. // cbits is the number of low order bits in cur that are valid. // sbits is the number of bits of the symbol prefix being decoded. cur, cbits, sbits := uint(0), uint8(0), uint8(0) for _, b := range v { cur = cur<<8 | uint(b) cbits += 8 sbits += 8 for cbits >= 8 { idx := byte(cur >> (cbits - 8)) n = n.children[idx] if n == nil {
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri Aug 05 19:54:32 UTC 2022 - 5.8K bytes - Viewed (0) -
src/runtime/mcheckmark.go
} // setCheckmark throws if marking object is a checkmarks violation, // and otherwise sets obj's checkmark. It returns true if obj was // already checkmarked. func setCheckmark(obj, base, off uintptr, mbits markBits) bool { if !mbits.isMarked() { printlock() print("runtime: checkmarks found unexpected unmarked object obj=", hex(obj), "\n") print("runtime: found obj at *(", hex(base), "+", hex(off), ")\n")
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 25 19:53:03 UTC 2024 - 2.8K bytes - Viewed (0) -
android/guava/src/com/google/common/util/concurrent/AtomicDoubleArray.java
* from {@link Double#equals}, as if implemented by: * * <pre>{@code * static boolean bitEquals(double x, double y) { * long xBits = Double.doubleToRawLongBits(x); * long yBits = Double.doubleToRawLongBits(y); * return xBits == yBits; * } * }</pre> * * @author Doug Lea * @author Martin Buchholz * @since 11.0 */ @GwtIncompatible @J2ktIncompatible
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Apr 04 09:45:04 UTC 2023 - 8K bytes - Viewed (0) -
src/hash/crc32/gen_const_ppc64le.go
} } func genCrc32ConstTable(w *bytes.Buffer, poly uint32, polyid string) { ref_poly := reflect_bits(uint64(poly), 32) fmt.Fprintf(w, "\n\t/* Reduce %d kbits to 1024 bits */\n", blocking*8) j := 0 for i := (blocking * 8) - 1024; i > 0; i -= 1024 { a := reflect_bits(get_remainder(ref_poly, 32, uint(i)), 32) << 1 b := reflect_bits(get_remainder(ref_poly, 32, uint(i+64)), 32) << 1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Feb 19 20:44:20 UTC 2024 - 4.5K bytes - Viewed (0) -
guava/src/com/google/common/util/concurrent/AtomicDouble.java
* from {@link Double#equals}, as if implemented by: * * <pre>{@code * static boolean bitEquals(double x, double y) { * long xBits = Double.doubleToRawLongBits(x); * long yBits = Double.doubleToRawLongBits(y); * return xBits == yBits; * } * }</pre> * * <p>It is possible to write a more scalable updater, at the cost of giving up strict atomicity. * See for example <a
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Tue Apr 04 09:45:04 UTC 2023 - 9.5K bytes - Viewed (0)