Search Options

Results per page
Sort
Preferred Languages
Advance

Results 31 - 40 of 130 for doffsets (0.13 sec)

  1. src/internal/zstd/zstd.go

    	// The number of bytes read from r up to the start of the current
    	// block, for error reporting.
    	blockOffset int64
    
    	// Buffered decompressed data.
    	buffer []byte
    	// Current read offset in buffer.
    	off int
    
    	// The current repeated offsets.
    	repeatedOffset1 uint32
    	repeatedOffset2 uint32
    	repeatedOffset3 uint32
    
    	// The current Huffman tree used for compressing literals.
    	huffmanTable     []uint16
    	huffmanTableBits int
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Jan 30 04:10:45 UTC 2024
    - 12.7K bytes
    - Viewed (0)
  2. src/encoding/xml/read_test.go

    not being told what column the text began in, so it
    didn't know where to put the tab stops.  Another problem
    was that some of the code assumed that string byte
    offsets were the same as column offsets, which is only
    true if there are no tabs.
    
    In the process of fixing this, I cleaned up the arguments
    to Fold and ExpandTabs and renamed them Break and
    _ExpandTabs so that I could be sure that I found all the
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Mar 26 19:58:28 UTC 2024
    - 29.1K bytes
    - Viewed (0)
  3. src/cmd/internal/dwarf/dwarf.go

    			}
    
    			// This slice will hold the offset in bytes for each child
    			// variable DIE with respect to the start of the parent
    			// subprogram DIE.
    			for _, v := range flattened {
    				offsets = append(offsets, int32(ctxt.CurrentOffset(s.Absfn)))
    				putAbstractVar(ctxt, s.Absfn, v)
    			}
    		}
    	}
    	ctxt.RecordChildDieOffsets(s.Absfn, flattened, offsets)
    
    	Uleb128put(ctxt, s.Absfn, 0)
    	return nil
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed Mar 06 15:23:18 UTC 2024
    - 43K bytes
    - Viewed (0)
  4. src/go/ast/print.go

    }
    
    // Fprint prints the (sub-)tree starting at AST node x to w.
    // If fset != nil, position information is interpreted relative
    // to that file set. Otherwise positions are printed as integer
    // values (file set specific offsets).
    //
    // A non-nil [FieldFilter] f may be provided to control the output:
    // struct fields for which f(fieldname, fieldvalue) is true are
    // printed; all others are filtered from the output. Unexported
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Mar 28 21:32:41 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  5. src/internal/coverage/encodemeta/encode.go

    	}
    	if err := binary.Write(w, binary.LittleEndian, mh); err != nil {
    		return digest, fmt.Errorf("error writing meta-file header: %v", err)
    	}
    	off := int64(coverage.CovMetaHeaderSize)
    
    	// Write function offsets section
    	off = b.emitFuncOffsets(w, off)
    
    	// Check for any errors up to this point.
    	if b.werr != nil {
    		return digest, b.werr
    	}
    
    	// Write string table.
    	if err := b.stab.Write(w); err != nil {
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu Feb 29 17:16:10 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  6. cmd/httprange.go

    // Case 2: bytes=1-10 (absolute start and end offsets) -> RangeSpec{false, 1, 10}
    // Case 3: bytes=10- (absolute start offset with end offset unspecified) -> RangeSpec{false, 10, -1}
    // Case 4: bytes=-30 (suffix length specification) -> RangeSpec{true, -30, -1}
    type HTTPRangeSpec struct {
    	// Does the range spec refer to a suffix of the object?
    	IsSuffixLength bool
    
    	// Start and end offset specified in range spec
    	Start, End int64
    }
    
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Thu May 09 08:44:07 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  7. cmd/erasure-decode_test.go

    	}
    
    	// To generate random offset/length.
    	r := rand.New(rand.NewSource(UTCNow().UnixNano()))
    
    	buf := &bytes.Buffer{}
    
    	// Verify erasure.Decode() for random offsets and lengths.
    	for i := 0; i < iterations; i++ {
    		offset := r.Int63n(length)
    		readLen := r.Int63n(length - offset)
    
    		expected := data[offset : offset+readLen]
    
    		// Get the checksums of the current part.
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Tue Jan 30 20:43:25 UTC 2024
    - 21.1K bytes
    - Viewed (0)
  8. src/internal/xcoff/file.go

    		return nil, err
    	}
    
    	offset := 0
    	// First import file ID is the default LIBPATH value
    	libpath := cstring(table[offset:])
    	f.LibraryPaths = strings.Split(libpath, ":")
    	offset += len(libpath) + 3 // 3 null bytes
    	all := make([]string, 0)
    	for i := 1; i < int(nimpid); i++ {
    		impidpath := cstring(table[offset:])
    		offset += len(impidpath) + 1
    		impidbase := cstring(table[offset:])
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Feb 12 14:42:29 UTC 2024
    - 17.3K bytes
    - Viewed (0)
  9. cmd/bitrot.go

    		if left < shardSize {
    			shardSize = left
    		}
    
    		read, err := io.CopyBuffer(h, io.LimitReader(r, shardSize), *bufp)
    		if err != nil {
    			// Read's failed for object with right size, at different offsets.
    			return errFileCorrupt
    		}
    
    		left -= read
    		if !bytes.Equal(h.Sum(nil), hashBuf[:n]) {
    			return errFileCorrupt
    		}
    	}
    	return nil
    }
    
    Registered: Sun Jun 16 00:44:34 UTC 2024
    - Last Modified: Tue Jan 30 20:43:25 UTC 2024
    - 7.6K bytes
    - Viewed (0)
  10. src/index/suffixarray/sais2.go

    		return
    	}
    	if len(text) == 1 {
    		sa[0] = 0
    		return
    	}
    
    	// Establish slices indexed by text character
    	// holding character frequency and bucket-sort offsets.
    	// If there's only enough tmp for one slice,
    	// we make it the bucket offsets and recompute
    	// the character frequency each time we need it.
    	var freq, bucket []int64
    	if len(tmp) >= 2*textMax {
    		freq, bucket = tmp[:textMax], tmp[textMax:2*textMax]
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Mon Mar 18 23:57:18 UTC 2024
    - 52.3K bytes
    - Viewed (0)
Back to top